diff --git a/CHANGELOG.md b/CHANGELOG.md index a875445d407..841bcba7bcc 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,56 @@ # Changelog +## [2.16.0](https://www.github.com/googleapis/google-api-python-client/compare/v2.15.0...v2.16.0) (2021-08-10) + + +### Features + +* **accesscontextmanager:** update the api https://github.com/googleapis/google-api-python-client/commit/30216a669249442cac8f0fb8bb347b1352d8f087 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **alertcenter:** update the api https://github.com/googleapis/google-api-python-client/commit/39b084706537111e8403be6e69f0fc9d82b2f383 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **androidpublisher:** update the api https://github.com/googleapis/google-api-python-client/commit/cf67afc22e94f856773895a4e603e7a9a6bfa20b ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **apigee:** update the api https://github.com/googleapis/google-api-python-client/commit/4485c5f3b32c9bda4f50a2a96c5870414f7d870f ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **apigee:** update the api https://github.com/googleapis/google-api-python-client/commit/995336984e11fb9f91308d14a68faf8f3091d1fa ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **appengine:** update the api https://github.com/googleapis/google-api-python-client/commit/eb7a571470cef08641224558a7bd8eaa07a41bad ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **bigquery:** update the api https://github.com/googleapis/google-api-python-client/commit/304bbde2360066caf55575e3be5a04fdc8bf8b09 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **chromemanagement:** update the api https://github.com/googleapis/google-api-python-client/commit/0ba28b47236a81a996a3607567b61ab38150617d ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **chromepolicy:** update the api https://github.com/googleapis/google-api-python-client/commit/5654776fdc8361aa0703a7dca8069b576a1b2f73 ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **cloudasset:** update the api https://github.com/googleapis/google-api-python-client/commit/792aa5593ea64ceb4b565e950e153e396274b3b8 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **cloudbuild:** update the api https://github.com/googleapis/google-api-python-client/commit/437e37f1c36268464f90e075ffeaef61580de237 ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **cloudbuild:** update the api https://github.com/googleapis/google-api-python-client/commit/6b06387ca29e76d26f257c7a4eb6864fe27e082e ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **cloudchannel:** update the api https://github.com/googleapis/google-api-python-client/commit/0b0444ea192f79c5564745be8b1d52b52a74d1fb ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **cloudkms:** update the api https://github.com/googleapis/google-api-python-client/commit/7dc278459cbd32bf15b39633327743cfa0beeea3 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **cloudkms:** update the api https://github.com/googleapis/google-api-python-client/commit/ebd3f49f78738792032e431b73233ae0c458bae3 ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **composer:** update the api https://github.com/googleapis/google-api-python-client/commit/92131bff38ab7644e549f3d71f9c0a84755455db ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **containeranalysis:** update the api https://github.com/googleapis/google-api-python-client/commit/9ab94f3f4c2ccafdf8b298dad8c6a5c2aa61a606 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **containeranalysis:** update the api https://github.com/googleapis/google-api-python-client/commit/d446928f941d858022f0e1a1911bbf185920159d ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **container:** update the api https://github.com/googleapis/google-api-python-client/commit/53d8b4b3e3c5d16ffd14ba1af1cf1769d9249067 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **dialogflow:** update the api https://github.com/googleapis/google-api-python-client/commit/0feb05616eb28db7c35e128ebbf338b63446b8cf ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **dialogflow:** update the api https://github.com/googleapis/google-api-python-client/commit/b49bfdaedbed3378b061e85f937a36e97732fcd4 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **iam:** update the api https://github.com/googleapis/google-api-python-client/commit/0832247d126965b472a271167028499f015de1ae ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **ideahub:** update the api https://github.com/googleapis/google-api-python-client/commit/24483a4f512922f809fc8352b9407e606856b0e2 ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **metastore:** update the api https://github.com/googleapis/google-api-python-client/commit/897beb3754da50e117292f5954265076804acb7f ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **metastore:** update the api https://github.com/googleapis/google-api-python-client/commit/dd83236343d603e964613b16e9afa25eff60f97c ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **monitoring:** update the api https://github.com/googleapis/google-api-python-client/commit/75a5ced2b372723c21d45b172dd69e0bb91c5509 ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **ondemandscanning:** update the api https://github.com/googleapis/google-api-python-client/commit/6ffbe182425ec217230fc083e217676e915bb786 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **osconfig:** update the api https://github.com/googleapis/google-api-python-client/commit/c8b511a21f7fc7f2471d5f7a3b2d3760e4f8a629 ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **osconfig:** update the api https://github.com/googleapis/google-api-python-client/commit/e079d43be4291ca10be7caf432012c6553e0398e ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **oslogin:** update the api https://github.com/googleapis/google-api-python-client/commit/d3ec653bdd293a63ae0b3772ce83e1fda73d5de5 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **pubsublite:** update the api https://github.com/googleapis/google-api-python-client/commit/4287a7d537741391a9afe9e669b98010ed4fc0ab ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **pubsub:** update the api https://github.com/googleapis/google-api-python-client/commit/06dfff22baec2551508b93e29e4c36fa442ab299 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **recaptchaenterprise:** update the api https://github.com/googleapis/google-api-python-client/commit/ebfeb8fc00a1c6a8603b35640845c5cdacf53cb2 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **recommender:** update the api https://github.com/googleapis/google-api-python-client/commit/01f2d6cc989ce337537a51ead8ffd3d6fc7e6c5d ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **speech:** update the api https://github.com/googleapis/google-api-python-client/commit/601afcf08fd96421b64ef4c6f098f09f0748ce69 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **speech:** update the api https://github.com/googleapis/google-api-python-client/commit/689fff21696add03b8c3ab843374b6bd2dd3cc16 ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **sqladmin:** update the api https://github.com/googleapis/google-api-python-client/commit/41d51e34759b181692ed96d9d490a9cfc5a28459 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **storagetransfer:** update the api https://github.com/googleapis/google-api-python-client/commit/24564836842f792e9373ea505d97e775f64a5960 ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) +* **tagmanager:** update the api https://github.com/googleapis/google-api-python-client/commit/47a522aac79ae9283a0c7ee7a2d0716e605d8c21 ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) +* **transcoder:** update the api https://github.com/googleapis/google-api-python-client/commit/1e0b0854e31f52013a8c5423efbd5e34c953e08c ([3e4b5db](https://www.github.com/googleapis/google-api-python-client/commit/3e4b5db2a4aeceb4b99f280e8a843c65284c2f9f)) + + +### Bug Fixes + +* **fcm:** update the api https://github.com/googleapis/google-api-python-client/commit/f1dd412cad2a2cdd1863bb2942cf07fc6a42b649 ([b539cc4](https://www.github.com/googleapis/google-api-python-client/commit/b539cc475e81dc6f0d3c0f75b1e0445bb79fe9e5)) + ## [2.15.0](https://www.github.com/googleapis/google-api-python-client/compare/v2.14.1...v2.15.0) (2021-07-27) diff --git a/docs/dyn/accesscontextmanager_v1.accessPolicies.accessLevels.html b/docs/dyn/accesscontextmanager_v1.accessPolicies.accessLevels.html index fd6808c05fc..46771be5bdb 100644 --- a/docs/dyn/accesscontextmanager_v1.accessPolicies.accessLevels.html +++ b/docs/dyn/accesscontextmanager_v1.accessPolicies.accessLevels.html @@ -161,7 +161,7 @@
Update an Access Level. The longrunning operation from this RPC will have a successful status once the changes to the Access Level have propagated to long-lasting storage. Access Levels containing errors will result in an error response for the first error encountered.
Args:
- name: string, Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/accessLevels/{short_name}`. The maximum length of the `short_name` component is 50 characters. (required)
+ name: string, Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/accessLevels/{access_level}`. The maximum length of the `access_level` component is 50 characters. (required)
body: object, The request body.
The object takes the form of:
@@ -451,7 +451,7 @@ Method Details
},
},
"description": "A String", # Description of the `AccessLevel` and its use. Does not affect behavior.
- "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/accessLevels/{short_name}`. The maximum length of the `short_name` component is 50 characters.
+ "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/accessLevels/{access_level}`. The maximum length of the `access_level` component is 50 characters.
"title": "A String", # Human readable title. Must be unique within the Policy.
}
@@ -544,7 +544,7 @@ Method Details
},
},
"description": "A String", # Description of the `AccessLevel` and its use. Does not affect behavior.
- "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/accessLevels/{short_name}`. The maximum length of the `short_name` component is 50 characters.
+ "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/accessLevels/{access_level}`. The maximum length of the `access_level` component is 50 characters.
"title": "A String", # Human readable title. Must be unique within the Policy.
},
],
diff --git a/docs/dyn/accesscontextmanager_v1.accessPolicies.html b/docs/dyn/accesscontextmanager_v1.accessPolicies.html
index 78dc880affb..30ca5772926 100644
--- a/docs/dyn/accesscontextmanager_v1.accessPolicies.html
+++ b/docs/dyn/accesscontextmanager_v1.accessPolicies.html
@@ -121,7 +121,7 @@ Method Details
{ # `AccessPolicy` is a container for `AccessLevels` (which define the necessary attributes to use Google Cloud services) and `ServicePerimeters` (which define regions of services able to freely pass data within a perimeter). An access policy is globally visible within an organization, and the restrictions it specifies apply to all projects within an organization.
"etag": "A String", # Output only. An opaque identifier for the current version of the `AccessPolicy`. This will always be a strongly validated etag, meaning that two Access Polices will be identical if and only if their etags are identical. Clients should not expect this to be in any specific format.
- "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{policy_id}`
+ "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{access_policy}`
"parent": "A String", # Required. The parent of this `AccessPolicy` in the Cloud Resource Hierarchy. Currently immutable once created. Format: `organizations/{organization_id}`
"title": "A String", # Required. Human readable title. Does not affect behavior.
}
@@ -206,7 +206,7 @@ Method Details
{ # `AccessPolicy` is a container for `AccessLevels` (which define the necessary attributes to use Google Cloud services) and `ServicePerimeters` (which define regions of services able to freely pass data within a perimeter). An access policy is globally visible within an organization, and the restrictions it specifies apply to all projects within an organization.
"etag": "A String", # Output only. An opaque identifier for the current version of the `AccessPolicy`. This will always be a strongly validated etag, meaning that two Access Polices will be identical if and only if their etags are identical. Clients should not expect this to be in any specific format.
- "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{policy_id}`
+ "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{access_policy}`
"parent": "A String", # Required. The parent of this `AccessPolicy` in the Cloud Resource Hierarchy. Currently immutable once created. Format: `organizations/{organization_id}`
"title": "A String", # Required. Human readable title. Does not affect behavior.
}
@@ -232,7 +232,7 @@ Update an AccessPolicy. The longrunning Operation from this RPC will have a successful status once the changes to the AccessPolicy have propagated to long-lasting storage. Syntactic and basic semantic errors will be returned in `metadata` as a BadRequest proto.
Args:
- name: string, Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{policy_id}` (required)
+ name: string, Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{access_policy}` (required)
body: object, The request body.
The object takes the form of:
{ # `AccessPolicy` is a container for `AccessLevels` (which define the necessary attributes to use Google Cloud services) and `ServicePerimeters` (which define regions of services able to freely pass data within a perimeter). An access policy is globally visible within an organization, and the restrictions it specifies apply to all projects within an organization.
"etag": "A String", # Output only. An opaque identifier for the current version of the `AccessPolicy`. This will always be a strongly validated etag, meaning that two Access Polices will be identical if and only if their etags are identical. Clients should not expect this to be in any specific format.
- "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{policy_id}`
+ "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{access_policy}`
"parent": "A String", # Required. The parent of this `AccessPolicy` in the Cloud Resource Hierarchy. Currently immutable once created. Format: `organizations/{organization_id}`
"title": "A String", # Required. Human readable title. Does not affect behavior.
}
diff --git a/docs/dyn/accesscontextmanager_v1.accessPolicies.servicePerimeters.html b/docs/dyn/accesscontextmanager_v1.accessPolicies.servicePerimeters.html
index e6a508ee4b2..e8b7f14c08b 100644
--- a/docs/dyn/accesscontextmanager_v1.accessPolicies.servicePerimeters.html
+++ b/docs/dyn/accesscontextmanager_v1.accessPolicies.servicePerimeters.html
@@ -160,7 +160,7 @@ Method Details
{ # `ServicePerimeter` describes a set of Google Cloud resources which can freely import and export data amongst themselves, but not export outside of the `ServicePerimeter`. If a request with a source within this `ServicePerimeter` has a target outside of the `ServicePerimeter`, the request will be blocked. Otherwise the request is allowed. There are two types of Service Perimeter - Regular and Bridge. Regular Service Perimeters cannot overlap, a single Google Cloud project can only belong to a single regular Service Perimeter. Service Perimeter Bridges can contain only Google Cloud projects as members, a single Google Cloud project may belong to multiple Service Perimeter Bridges.
"description": "A String", # Description of the `ServicePerimeter` and its use. Does not affect behavior.
- "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}`
+ "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}`
"perimeterType": "A String", # Perimeter type indicator. A single project is allowed to be a member of single regular perimeter, but multiple service perimeter bridges. A project cannot be a included in a perimeter bridge without being included in regular perimeter. For perimeter bridges, the restricted service list as well as access level lists must be empty.
"spec": { # `ServicePerimeterConfig` specifies a set of Google Cloud resources that describe specific Service Perimeter configuration. # Proposed (or dry run) ServicePerimeter configuration. This configuration allows to specify and test ServicePerimeter configuration without enforcing actual access restrictions. Only allowed to be set when the "use_explicit_dry_run_spec" flag is set.
"accessLevels": [ # A list of `AccessLevel` resource names that allow resources within the `ServicePerimeter` to be accessed from the internet. `AccessLevels` listed must be in the same policy as this `ServicePerimeter`. Referencing a nonexistent `AccessLevel` is a syntax error. If no `AccessLevel` names are listed, resources within the perimeter can only be accessed via Google Cloud calls with request origins within the perimeter. Example: `"accessPolicies/MY_POLICY/accessLevels/MY_LEVEL"`. For Service Perimeter Bridge, must be empty.
@@ -396,7 +396,7 @@ Method Details
{ # `ServicePerimeter` describes a set of Google Cloud resources which can freely import and export data amongst themselves, but not export outside of the `ServicePerimeter`. If a request with a source within this `ServicePerimeter` has a target outside of the `ServicePerimeter`, the request will be blocked. Otherwise the request is allowed. There are two types of Service Perimeter - Regular and Bridge. Regular Service Perimeters cannot overlap, a single Google Cloud project can only belong to a single regular Service Perimeter. Service Perimeter Bridges can contain only Google Cloud projects as members, a single Google Cloud project may belong to multiple Service Perimeter Bridges.
"description": "A String", # Description of the `ServicePerimeter` and its use. Does not affect behavior.
- "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}`
+ "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}`
"perimeterType": "A String", # Perimeter type indicator. A single project is allowed to be a member of single regular perimeter, but multiple service perimeter bridges. A project cannot be a included in a perimeter bridge without being included in regular perimeter. For perimeter bridges, the restricted service list as well as access level lists must be empty.
"spec": { # `ServicePerimeterConfig` specifies a set of Google Cloud resources that describe specific Service Perimeter configuration. # Proposed (or dry run) ServicePerimeter configuration. This configuration allows to specify and test ServicePerimeter configuration without enforcing actual access restrictions. Only allowed to be set when the "use_explicit_dry_run_spec" flag is set.
"accessLevels": [ # A list of `AccessLevel` resource names that allow resources within the `ServicePerimeter` to be accessed from the internet. `AccessLevels` listed must be in the same policy as this `ServicePerimeter`. Referencing a nonexistent `AccessLevel` is a syntax error. If no `AccessLevel` names are listed, resources within the perimeter can only be accessed via Google Cloud calls with request origins within the perimeter. Example: `"accessPolicies/MY_POLICY/accessLevels/MY_LEVEL"`. For Service Perimeter Bridge, must be empty.
@@ -574,7 +574,7 @@ Method Details
"servicePerimeters": [ # List of the Service Perimeter instances.
{ # `ServicePerimeter` describes a set of Google Cloud resources which can freely import and export data amongst themselves, but not export outside of the `ServicePerimeter`. If a request with a source within this `ServicePerimeter` has a target outside of the `ServicePerimeter`, the request will be blocked. Otherwise the request is allowed. There are two types of Service Perimeter - Regular and Bridge. Regular Service Perimeters cannot overlap, a single Google Cloud project can only belong to a single regular Service Perimeter. Service Perimeter Bridges can contain only Google Cloud projects as members, a single Google Cloud project may belong to multiple Service Perimeter Bridges.
"description": "A String", # Description of the `ServicePerimeter` and its use. Does not affect behavior.
- "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}`
+ "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}`
"perimeterType": "A String", # Perimeter type indicator. A single project is allowed to be a member of single regular perimeter, but multiple service perimeter bridges. A project cannot be a included in a perimeter bridge without being included in regular perimeter. For perimeter bridges, the restricted service list as well as access level lists must be empty.
"spec": { # `ServicePerimeterConfig` specifies a set of Google Cloud resources that describe specific Service Perimeter configuration. # Proposed (or dry run) ServicePerimeter configuration. This configuration allows to specify and test ServicePerimeter configuration without enforcing actual access restrictions. Only allowed to be set when the "use_explicit_dry_run_spec" flag is set.
"accessLevels": [ # A list of `AccessLevel` resource names that allow resources within the `ServicePerimeter` to be accessed from the internet. `AccessLevels` listed must be in the same policy as this `ServicePerimeter`. Referencing a nonexistent `AccessLevel` is a syntax error. If no `AccessLevel` names are listed, resources within the perimeter can only be accessed via Google Cloud calls with request origins within the perimeter. Example: `"accessPolicies/MY_POLICY/accessLevels/MY_LEVEL"`. For Service Perimeter Bridge, must be empty.
@@ -752,13 +752,13 @@ Method Details
Update a Service Perimeter. The longrunning operation from this RPC will have a successful status once the changes to the Service Perimeter have propagated to long-lasting storage. Service Perimeter containing errors will result in an error response for the first error encountered.
Args:
- name: string, Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}` (required)
+ name: string, Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}` (required)
body: object, The request body.
The object takes the form of:
{ # `ServicePerimeter` describes a set of Google Cloud resources which can freely import and export data amongst themselves, but not export outside of the `ServicePerimeter`. If a request with a source within this `ServicePerimeter` has a target outside of the `ServicePerimeter`, the request will be blocked. Otherwise the request is allowed. There are two types of Service Perimeter - Regular and Bridge. Regular Service Perimeters cannot overlap, a single Google Cloud project can only belong to a single regular Service Perimeter. Service Perimeter Bridges can contain only Google Cloud projects as members, a single Google Cloud project may belong to multiple Service Perimeter Bridges.
"description": "A String", # Description of the `ServicePerimeter` and its use. Does not affect behavior.
- "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}`
+ "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}`
"perimeterType": "A String", # Perimeter type indicator. A single project is allowed to be a member of single regular perimeter, but multiple service perimeter bridges. A project cannot be a included in a perimeter bridge without being included in regular perimeter. For perimeter bridges, the restricted service list as well as access level lists must be empty.
"spec": { # `ServicePerimeterConfig` specifies a set of Google Cloud resources that describe specific Service Perimeter configuration. # Proposed (or dry run) ServicePerimeter configuration. This configuration allows to specify and test ServicePerimeter configuration without enforcing actual access restrictions. Only allowed to be set when the "use_explicit_dry_run_spec" flag is set.
"accessLevels": [ # A list of `AccessLevel` resource names that allow resources within the `ServicePerimeter` to be accessed from the internet. `AccessLevels` listed must be in the same policy as this `ServicePerimeter`. Referencing a nonexistent `AccessLevel` is a syntax error. If no `AccessLevel` names are listed, resources within the perimeter can only be accessed via Google Cloud calls with request origins within the perimeter. Example: `"accessPolicies/MY_POLICY/accessLevels/MY_LEVEL"`. For Service Perimeter Bridge, must be empty.
@@ -958,7 +958,7 @@ Method Details
"servicePerimeters": [ # Required. The desired Service Perimeters that should replace all existing Service Perimeters in the Access Policy.
{ # `ServicePerimeter` describes a set of Google Cloud resources which can freely import and export data amongst themselves, but not export outside of the `ServicePerimeter`. If a request with a source within this `ServicePerimeter` has a target outside of the `ServicePerimeter`, the request will be blocked. Otherwise the request is allowed. There are two types of Service Perimeter - Regular and Bridge. Regular Service Perimeters cannot overlap, a single Google Cloud project can only belong to a single regular Service Perimeter. Service Perimeter Bridges can contain only Google Cloud projects as members, a single Google Cloud project may belong to multiple Service Perimeter Bridges.
"description": "A String", # Description of the `ServicePerimeter` and its use. Does not affect behavior.
- "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}`
+ "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}`
"perimeterType": "A String", # Perimeter type indicator. A single project is allowed to be a member of single regular perimeter, but multiple service perimeter bridges. A project cannot be a included in a perimeter bridge without being included in regular perimeter. For perimeter bridges, the restricted service list as well as access level lists must be empty.
"spec": { # `ServicePerimeterConfig` specifies a set of Google Cloud resources that describe specific Service Perimeter configuration. # Proposed (or dry run) ServicePerimeter configuration. This configuration allows to specify and test ServicePerimeter configuration without enforcing actual access restrictions. Only allowed to be set when the "use_explicit_dry_run_spec" flag is set.
"accessLevels": [ # A list of `AccessLevel` resource names that allow resources within the `ServicePerimeter` to be accessed from the internet. `AccessLevels` listed must be in the same policy as this `ServicePerimeter`. Referencing a nonexistent `AccessLevel` is a syntax error. If no `AccessLevel` names are listed, resources within the perimeter can only be accessed via Google Cloud calls with request origins within the perimeter. Example: `"accessPolicies/MY_POLICY/accessLevels/MY_LEVEL"`. For Service Perimeter Bridge, must be empty.
diff --git a/docs/dyn/alertcenter_v1beta1.alerts.html b/docs/dyn/alertcenter_v1beta1.alerts.html
index c8844a1c916..233404da016 100644
--- a/docs/dyn/alertcenter_v1beta1.alerts.html
+++ b/docs/dyn/alertcenter_v1beta1.alerts.html
@@ -248,7 +248,7 @@ Method Details
"updateTime": "A String", # Output only. The time this metadata was last updated.
},
"securityInvestigationToolLink": "A String", # Output only. An optional [Security Investigation Tool](https://support.google.com/a/answer/7575955) query for this alert.
- "source": "A String", # Required. A unique identifier for the system that reported the alert. This is output only after alert is created. Supported sources are any of the following: * Google Operations * Mobile device management * Gmail phishing * Data Loss Prevention * Domain wide takeout * State sponsored attack * Google identity
+ "source": "A String", # Required. A unique identifier for the system that reported the alert. This is output only after alert is created. Supported sources are any of the following: * Google Operations * Mobile device management * Gmail phishing * Data Loss Prevention * Domain wide takeout * State sponsored attack * Google identity * Apps outage
"startTime": "A String", # Required. The time the event that caused this alert was started or detected.
"type": "A String", # Required. The type of the alert. This is output only after alert is created. For a list of available alert types see [Google Workspace Alert types](https://developers.google.com/admin-sdk/alertcenter/reference/alert-types).
"updateTime": "A String", # Output only. The time this alert was last updated.
@@ -321,7 +321,7 @@ Method Details
"updateTime": "A String", # Output only. The time this metadata was last updated.
},
"securityInvestigationToolLink": "A String", # Output only. An optional [Security Investigation Tool](https://support.google.com/a/answer/7575955) query for this alert.
- "source": "A String", # Required. A unique identifier for the system that reported the alert. This is output only after alert is created. Supported sources are any of the following: * Google Operations * Mobile device management * Gmail phishing * Data Loss Prevention * Domain wide takeout * State sponsored attack * Google identity
+ "source": "A String", # Required. A unique identifier for the system that reported the alert. This is output only after alert is created. Supported sources are any of the following: * Google Operations * Mobile device management * Gmail phishing * Data Loss Prevention * Domain wide takeout * State sponsored attack * Google identity * Apps outage
"startTime": "A String", # Required. The time the event that caused this alert was started or detected.
"type": "A String", # Required. The type of the alert. This is output only after alert is created. For a list of available alert types see [Google Workspace Alert types](https://developers.google.com/admin-sdk/alertcenter/reference/alert-types).
"updateTime": "A String", # Output only. The time this alert was last updated.
@@ -386,7 +386,7 @@ Method Details
"updateTime": "A String", # Output only. The time this metadata was last updated.
},
"securityInvestigationToolLink": "A String", # Output only. An optional [Security Investigation Tool](https://support.google.com/a/answer/7575955) query for this alert.
- "source": "A String", # Required. A unique identifier for the system that reported the alert. This is output only after alert is created. Supported sources are any of the following: * Google Operations * Mobile device management * Gmail phishing * Data Loss Prevention * Domain wide takeout * State sponsored attack * Google identity
+ "source": "A String", # Required. A unique identifier for the system that reported the alert. This is output only after alert is created. Supported sources are any of the following: * Google Operations * Mobile device management * Gmail phishing * Data Loss Prevention * Domain wide takeout * State sponsored attack * Google identity * Apps outage
"startTime": "A String", # Required. The time the event that caused this alert was started or detected.
"type": "A String", # Required. The type of the alert. This is output only after alert is created. For a list of available alert types see [Google Workspace Alert types](https://developers.google.com/admin-sdk/alertcenter/reference/alert-types).
"updateTime": "A String", # Output only. The time this alert was last updated.
diff --git a/docs/dyn/androidenterprise_v1.devices.html b/docs/dyn/androidenterprise_v1.devices.html
index eeddbab3ae3..a110b0b85da 100644
--- a/docs/dyn/androidenterprise_v1.devices.html
+++ b/docs/dyn/androidenterprise_v1.devices.html
@@ -94,7 +94,7 @@ Instance Methods
Sets whether a device's access to Google services is enabled or disabled. The device state takes effect only if enforcing EMM policies on Android devices is enabled in the Google Admin Console. Otherwise, the device state is ignored and all devices are allowed access to Google services. This is only supported for Google-managed users.
update(enterpriseId, userId, deviceId, body=None, updateMask=None, x__xgafv=None)
-Updates the device policy. To ensure the policy is properly enforced, you need to prevent unmanaged accounts from accessing Google Play by setting the allowed_accounts in the managed configuration for the Google Play package. See restrict accounts in Google Play.
+Updates the device policy. To ensure the policy is properly enforced, you need to prevent unmanaged accounts from accessing Google Play by setting the allowed_accounts in the managed configuration for the Google Play package. See restrict accounts in Google Play. When provisioning a new device, you should set the device policy using this method before adding the managed Google Play Account to the device, otherwise the policy will not be applied for a short period of time after adding the account to the device.
Method Details
close()
@@ -386,7 +386,7 @@ Method Details
update(enterpriseId, userId, deviceId, body=None, updateMask=None, x__xgafv=None)
- Updates the device policy. To ensure the policy is properly enforced, you need to prevent unmanaged accounts from accessing Google Play by setting the allowed_accounts in the managed configuration for the Google Play package. See restrict accounts in Google Play.
+ Updates the device policy. To ensure the policy is properly enforced, you need to prevent unmanaged accounts from accessing Google Play by setting the allowed_accounts in the managed configuration for the Google Play package. See restrict accounts in Google Play. When provisioning a new device, you should set the device policy using this method before adding the managed Google Play Account to the device, otherwise the policy will not be applied for a short period of time after adding the account to the device.
Args:
enterpriseId: string, The ID of the enterprise. (required)
diff --git a/docs/dyn/androidmanagement_v1.enterprises.devices.html b/docs/dyn/androidmanagement_v1.enterprises.devices.html
index 404be39f4c8..c67de4a08ec 100644
--- a/docs/dyn/androidmanagement_v1.enterprises.devices.html
+++ b/docs/dyn/androidmanagement_v1.enterprises.devices.html
@@ -295,7 +295,7 @@ Method Details
"eventType": "A String", # Event type.
},
],
- "memoryInfo": { # Information about device memory and storage. # Memory information. This information is only available if memoryInfoEnabled is true in the device's policy.
+ "memoryInfo": { # Information about device memory and storage. # Memory information: contains information about device memory and storage.
"totalInternalStorage": "A String", # Total internal storage on device in bytes.
"totalRam": "A String", # Total RAM on device in bytes.
},
@@ -597,7 +597,7 @@ Method Details
"eventType": "A String", # Event type.
},
],
- "memoryInfo": { # Information about device memory and storage. # Memory information. This information is only available if memoryInfoEnabled is true in the device's policy.
+ "memoryInfo": { # Information about device memory and storage. # Memory information: contains information about device memory and storage.
"totalInternalStorage": "A String", # Total internal storage on device in bytes.
"totalRam": "A String", # Total RAM on device in bytes.
},
@@ -857,7 +857,7 @@ Method Details
"eventType": "A String", # Event type.
},
],
- "memoryInfo": { # Information about device memory and storage. # Memory information. This information is only available if memoryInfoEnabled is true in the device's policy.
+ "memoryInfo": { # Information about device memory and storage. # Memory information: contains information about device memory and storage.
"totalInternalStorage": "A String", # Total internal storage on device in bytes.
"totalRam": "A String", # Total RAM on device in bytes.
},
@@ -1099,7 +1099,7 @@ Method Details
"eventType": "A String", # Event type.
},
],
- "memoryInfo": { # Information about device memory and storage. # Memory information. This information is only available if memoryInfoEnabled is true in the device's policy.
+ "memoryInfo": { # Information about device memory and storage. # Memory information: contains information about device memory and storage.
"totalInternalStorage": "A String", # Total internal storage on device in bytes.
"totalRam": "A String", # Total RAM on device in bytes.
},
diff --git a/docs/dyn/androidmanagement_v1.enterprises.policies.html b/docs/dyn/androidmanagement_v1.enterprises.policies.html
index 6b16fe0caf8..3d2e47236e7 100644
--- a/docs/dyn/androidmanagement_v1.enterprises.policies.html
+++ b/docs/dyn/androidmanagement_v1.enterprises.policies.html
@@ -422,7 +422,7 @@ Method Details
"deviceSettingsEnabled": True or False, # Whether device settings reporting is enabled.
"displayInfoEnabled": True or False, # Whether displays reporting is enabled. Report data is not available for personally owned devices with work profiles.
"hardwareStatusEnabled": True or False, # Whether hardware status reporting is enabled. Report data is not available for personally owned devices with work profiles.
- "memoryInfoEnabled": True or False, # Whether memory reporting is enabled.
+ "memoryInfoEnabled": True or False, # Whether memory event reporting is enabled.
"networkInfoEnabled": True or False, # Whether network info reporting is enabled.
"powerManagementEventsEnabled": True or False, # Whether power management event reporting is enabled. Report data is not available for personally owned devices with work profiles.
"softwareInfoEnabled": True or False, # Whether software info reporting is enabled.
@@ -773,7 +773,7 @@ Method Details
"deviceSettingsEnabled": True or False, # Whether device settings reporting is enabled.
"displayInfoEnabled": True or False, # Whether displays reporting is enabled. Report data is not available for personally owned devices with work profiles.
"hardwareStatusEnabled": True or False, # Whether hardware status reporting is enabled. Report data is not available for personally owned devices with work profiles.
- "memoryInfoEnabled": True or False, # Whether memory reporting is enabled.
+ "memoryInfoEnabled": True or False, # Whether memory event reporting is enabled.
"networkInfoEnabled": True or False, # Whether network info reporting is enabled.
"powerManagementEventsEnabled": True or False, # Whether power management event reporting is enabled. Report data is not available for personally owned devices with work profiles.
"softwareInfoEnabled": True or False, # Whether software info reporting is enabled.
@@ -1130,7 +1130,7 @@ Method Details
"deviceSettingsEnabled": True or False, # Whether device settings reporting is enabled.
"displayInfoEnabled": True or False, # Whether displays reporting is enabled. Report data is not available for personally owned devices with work profiles.
"hardwareStatusEnabled": True or False, # Whether hardware status reporting is enabled. Report data is not available for personally owned devices with work profiles.
- "memoryInfoEnabled": True or False, # Whether memory reporting is enabled.
+ "memoryInfoEnabled": True or False, # Whether memory event reporting is enabled.
"networkInfoEnabled": True or False, # Whether network info reporting is enabled.
"powerManagementEventsEnabled": True or False, # Whether power management event reporting is enabled. Report data is not available for personally owned devices with work profiles.
"softwareInfoEnabled": True or False, # Whether software info reporting is enabled.
@@ -1470,7 +1470,7 @@ Method Details
"deviceSettingsEnabled": True or False, # Whether device settings reporting is enabled.
"displayInfoEnabled": True or False, # Whether displays reporting is enabled. Report data is not available for personally owned devices with work profiles.
"hardwareStatusEnabled": True or False, # Whether hardware status reporting is enabled. Report data is not available for personally owned devices with work profiles.
- "memoryInfoEnabled": True or False, # Whether memory reporting is enabled.
+ "memoryInfoEnabled": True or False, # Whether memory event reporting is enabled.
"networkInfoEnabled": True or False, # Whether network info reporting is enabled.
"powerManagementEventsEnabled": True or False, # Whether power management event reporting is enabled. Report data is not available for personally owned devices with work profiles.
"softwareInfoEnabled": True or False, # Whether software info reporting is enabled.
diff --git a/docs/dyn/androidpublisher_v3.inappproducts.html b/docs/dyn/androidpublisher_v3.inappproducts.html
index 2fc53643373..8610b52e903 100644
--- a/docs/dyn/androidpublisher_v3.inappproducts.html
+++ b/docs/dyn/androidpublisher_v3.inappproducts.html
@@ -93,7 +93,7 @@ Instance Methods
patch(packageName, sku, autoConvertMissingPrices=None, body=None, x__xgafv=None)
Patches an in-app product (i.e. a managed product or a subscriptions).
- update(packageName, sku, autoConvertMissingPrices=None, body=None, x__xgafv=None)
+ update(packageName, sku, allowMissing=None, autoConvertMissingPrices=None, body=None, x__xgafv=None)
Updates an in-app product (i.e. a managed product or a subscriptions).
Method Details
@@ -383,7 +383,7 @@ Method Details
- update(packageName, sku, autoConvertMissingPrices=None, body=None, x__xgafv=None)
+ update(packageName, sku, allowMissing=None, autoConvertMissingPrices=None, body=None, x__xgafv=None)
Updates an in-app product (i.e. a managed product or a subscriptions).
Args:
@@ -422,6 +422,7 @@ Method Details
"trialPeriod": "A String", # Trial period, specified in ISO 8601 format. Acceptable values are anything between P7D (seven days) and P999D (999 days).
}
+ allowMissing: boolean, If set to true, and the in-app product with the given package_name and sku doesn't exist, the in-app product will be created.
autoConvertMissingPrices: boolean, If true the prices for all regions targeted by the parent app that don't have a price specified for this in-app product will be auto converted to the target currency based on the default price. Defaults to false.
x__xgafv: string, V1 error format.
Allowed values
diff --git a/docs/dyn/apigee_v1.organizations.apis.deployments.html b/docs/dyn/apigee_v1.organizations.apis.deployments.html
index 88002456282..9620cdd62d1 100644
--- a/docs/dyn/apigee_v1.organizations.apis.deployments.html
+++ b/docs/dyn/apigee_v1.organizations.apis.deployments.html
@@ -162,6 +162,7 @@ Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists.
},
],
+ "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`.
"state": "A String", # Current state of the deployment. This field is not populated in List APIs.
},
],
diff --git a/docs/dyn/apigee_v1.organizations.apis.html b/docs/dyn/apigee_v1.organizations.apis.html
index c9dfc7dcd79..88378852297 100644
--- a/docs/dyn/apigee_v1.organizations.apis.html
+++ b/docs/dyn/apigee_v1.organizations.apis.html
@@ -213,14 +213,17 @@ Method Details
An object of the form:
{ # Metadata describing the API proxy
- "latestRevisionId": "A String", # The id of the most recently created revision for this api proxy.
- "metaData": { # Metadata common to many entities in this API. # Metadata describing the API proxy.
+ "labels": { # User labels applied to this API Proxy.
+ "a_key": "A String",
+ },
+ "latestRevisionId": "A String", # Output only. The id of the most recently created revision for this api proxy.
+ "metaData": { # Metadata common to many entities in this API. # Output only. Metadata describing the API proxy.
"createdAt": "A String", # Time at which the API proxy was created, in milliseconds since epoch.
"lastModifiedAt": "A String", # Time at which the API proxy was most recently modified, in milliseconds since epoch.
"subType": "A String", # The type of entity described
},
- "name": "A String", # Name of the API proxy.
- "revision": [ # List of revisons defined for the API proxy.
+ "name": "A String", # Output only. Name of the API proxy.
+ "revision": [ # Output only. List of revisons defined for the API proxy.
"A String",
],
}
@@ -241,14 +244,17 @@ Method Details
An object of the form:
{ # Metadata describing the API proxy
- "latestRevisionId": "A String", # The id of the most recently created revision for this api proxy.
- "metaData": { # Metadata common to many entities in this API. # Metadata describing the API proxy.
+ "labels": { # User labels applied to this API Proxy.
+ "a_key": "A String",
+ },
+ "latestRevisionId": "A String", # Output only. The id of the most recently created revision for this api proxy.
+ "metaData": { # Metadata common to many entities in this API. # Output only. Metadata describing the API proxy.
"createdAt": "A String", # Time at which the API proxy was created, in milliseconds since epoch.
"lastModifiedAt": "A String", # Time at which the API proxy was most recently modified, in milliseconds since epoch.
"subType": "A String", # The type of entity described
},
- "name": "A String", # Name of the API proxy.
- "revision": [ # List of revisons defined for the API proxy.
+ "name": "A String", # Output only. Name of the API proxy.
+ "revision": [ # Output only. List of revisons defined for the API proxy.
"A String",
],
}
@@ -273,14 +279,17 @@ Method Details
{
"proxies": [
{ # Metadata describing the API proxy
- "latestRevisionId": "A String", # The id of the most recently created revision for this api proxy.
- "metaData": { # Metadata common to many entities in this API. # Metadata describing the API proxy.
+ "labels": { # User labels applied to this API Proxy.
+ "a_key": "A String",
+ },
+ "latestRevisionId": "A String", # Output only. The id of the most recently created revision for this api proxy.
+ "metaData": { # Metadata common to many entities in this API. # Output only. Metadata describing the API proxy.
"createdAt": "A String", # Time at which the API proxy was created, in milliseconds since epoch.
"lastModifiedAt": "A String", # Time at which the API proxy was most recently modified, in milliseconds since epoch.
"subType": "A String", # The type of entity described
},
- "name": "A String", # Name of the API proxy.
- "revision": [ # List of revisons defined for the API proxy.
+ "name": "A String", # Output only. Name of the API proxy.
+ "revision": [ # Output only. List of revisons defined for the API proxy.
"A String",
],
},
diff --git a/docs/dyn/apigee_v1.organizations.apis.revisions.deployments.html b/docs/dyn/apigee_v1.organizations.apis.revisions.deployments.html
index 1e285355213..702cef0bf3c 100644
--- a/docs/dyn/apigee_v1.organizations.apis.revisions.deployments.html
+++ b/docs/dyn/apigee_v1.organizations.apis.revisions.deployments.html
@@ -162,6 +162,7 @@ Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists.
},
],
+ "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`.
"state": "A String", # Current state of the deployment. This field is not populated in List APIs.
},
],
diff --git a/docs/dyn/apigee_v1.organizations.deployments.html b/docs/dyn/apigee_v1.organizations.deployments.html
index 0ae570e926b..4ce3a3d76f3 100644
--- a/docs/dyn/apigee_v1.organizations.deployments.html
+++ b/docs/dyn/apigee_v1.organizations.deployments.html
@@ -163,6 +163,7 @@ Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists.
},
],
+ "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`.
"state": "A String", # Current state of the deployment. This field is not populated in List APIs.
},
],
diff --git a/docs/dyn/apigee_v1.organizations.environments.apis.deployments.html b/docs/dyn/apigee_v1.organizations.environments.apis.deployments.html
index 72b8196751c..d96f2a03967 100644
--- a/docs/dyn/apigee_v1.organizations.environments.apis.deployments.html
+++ b/docs/dyn/apigee_v1.organizations.environments.apis.deployments.html
@@ -162,6 +162,7 @@ Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists.
},
],
+ "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`.
"state": "A String", # Current state of the deployment. This field is not populated in List APIs.
},
],
diff --git a/docs/dyn/apigee_v1.organizations.environments.apis.revisions.html b/docs/dyn/apigee_v1.organizations.environments.apis.revisions.html
index 09a7e0b8a0c..8484bad09b6 100644
--- a/docs/dyn/apigee_v1.organizations.environments.apis.revisions.html
+++ b/docs/dyn/apigee_v1.organizations.environments.apis.revisions.html
@@ -88,7 +88,7 @@ Instance Methods
close()
Close httplib2 connections.
- deploy(name, override=None, sequencedRollout=None, x__xgafv=None)
+ deploy(name, override=None, sequencedRollout=None, serviceAccount=None, x__xgafv=None)
Deploys a revision of an API proxy. If another revision of the same API proxy revision is currently deployed, set the `override` parameter to `true` to have this revision replace the currently deployed revision. You cannot invoke an API proxy until it has been deployed to an environment. After you deploy an API proxy revision, you cannot edit it. To edit the API proxy, you must create and deploy a new revision. For a request path `organizations/{org}/environments/{env}/apis/{api}/revisions/{rev}/deployments`, two permissions are required: * `apigee.deployments.create` on the resource `organizations/{org}/environments/{env}` * `apigee.proxyrevisions.deploy` on the resource `organizations/{org}/apis/{api}/revisions/{rev}`
getDeployments(name, x__xgafv=None)
@@ -103,13 +103,14 @@ Method Details
- deploy(name, override=None, sequencedRollout=None, x__xgafv=None)
+ deploy(name, override=None, sequencedRollout=None, serviceAccount=None, x__xgafv=None)
Deploys a revision of an API proxy. If another revision of the same API proxy revision is currently deployed, set the `override` parameter to `true` to have this revision replace the currently deployed revision. You cannot invoke an API proxy until it has been deployed to an environment. After you deploy an API proxy revision, you cannot edit it. To edit the API proxy, you must create and deploy a new revision. For a request path `organizations/{org}/environments/{env}/apis/{api}/revisions/{rev}/deployments`, two permissions are required: * `apigee.deployments.create` on the resource `organizations/{org}/environments/{env}` * `apigee.proxyrevisions.deploy` on the resource `organizations/{org}/apis/{api}/revisions/{rev}`
Args:
name: string, Required. Name of the API proxy revision deployment in the following format: `organizations/{org}/environments/{env}/apis/{api}/revisions/{rev}` (required)
override: boolean, Flag that specifies whether the new deployment replaces other deployed revisions of the API proxy in the environment. Set `override` to `true` to replace other deployed revisions. By default, `override` is `false` and the deployment is rejected if other revisions of the API proxy are deployed in the environment.
sequencedRollout: boolean, Flag that specifies whether to enable sequenced rollout. If set to `true`, a best-effort attempt will be made to roll out the routing rules corresponding to this deployment and the environment changes to add this deployment in a safe order. This reduces the risk of downtime that could be caused by changing the environment group's routing before the new destination for the affected traffic is ready to receive it. This should only be necessary if the new deployment will be capturing traffic from another environment under a shared environment group or if traffic will be rerouted to a different environment due to a base path removal. The [GenerateDeployChangeReport API](GenerateDeployChangeReport) may be used to examine routing changes before issuing the deployment request, and its response will indicate if a sequenced rollout is recommended for the deployment.
+ serviceAccount: string, Google Cloud IAM service account. The service account represents the identity of the deployed proxy, and determines what permissions it has. The format must be `{ACCOUNT_ID}@{PROJECT}.iam.gserviceaccount.com`.
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -178,6 +179,7 @@ Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists.
},
],
+ "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`.
"state": "A String", # Current state of the deployment. This field is not populated in List APIs.
}
@@ -256,6 +258,7 @@ Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists.
},
],
+ "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`.
"state": "A String", # Current state of the deployment. This field is not populated in List APIs.
}
diff --git a/docs/dyn/apigee_v1.organizations.environments.deployments.html b/docs/dyn/apigee_v1.organizations.environments.deployments.html
index 0a6d7b92f17..110b0969a0a 100644
--- a/docs/dyn/apigee_v1.organizations.environments.deployments.html
+++ b/docs/dyn/apigee_v1.organizations.environments.deployments.html
@@ -163,6 +163,7 @@ Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists.
},
],
+ "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`.
"state": "A String", # Current state of the deployment. This field is not populated in List APIs.
},
],
diff --git a/docs/dyn/apigee_v1.organizations.environments.sharedflows.deployments.html b/docs/dyn/apigee_v1.organizations.environments.sharedflows.deployments.html
index 74a70e93177..d22315312a3 100644
--- a/docs/dyn/apigee_v1.organizations.environments.sharedflows.deployments.html
+++ b/docs/dyn/apigee_v1.organizations.environments.sharedflows.deployments.html
@@ -162,6 +162,7 @@ Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists.
},
],
+ "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`.
"state": "A String", # Current state of the deployment. This field is not populated in List APIs.
},
],
diff --git a/docs/dyn/apigee_v1.organizations.environments.sharedflows.revisions.html b/docs/dyn/apigee_v1.organizations.environments.sharedflows.revisions.html
index ffa5efa2656..23ef5bd7b31 100644
--- a/docs/dyn/apigee_v1.organizations.environments.sharedflows.revisions.html
+++ b/docs/dyn/apigee_v1.organizations.environments.sharedflows.revisions.html
@@ -78,7 +78,7 @@ Instance Methods
close()
Close httplib2 connections.
- deploy(name, override=None, x__xgafv=None)
+ deploy(name, override=None, serviceAccount=None, x__xgafv=None)
Deploys a revision of a shared flow. If another revision of the same shared flow is currently deployed, set the `override` parameter to `true` to have this revision replace the currently deployed revision. You cannot use a shared flow until it has been deployed to an environment. For a request path `organizations/{org}/environments/{env}/sharedflows/{sf}/revisions/{rev}/deployments`, two permissions are required: * `apigee.deployments.create` on the resource `organizations/{org}/environments/{env}` * `apigee.sharedflowrevisions.deploy` on the resource `organizations/{org}/sharedflows/{sf}/revisions/{rev}`
getDeployments(name, x__xgafv=None)
@@ -93,12 +93,13 @@ Method Details
- deploy(name, override=None, x__xgafv=None)
+ deploy(name, override=None, serviceAccount=None, x__xgafv=None)
Deploys a revision of a shared flow. If another revision of the same shared flow is currently deployed, set the `override` parameter to `true` to have this revision replace the currently deployed revision. You cannot use a shared flow until it has been deployed to an environment. For a request path `organizations/{org}/environments/{env}/sharedflows/{sf}/revisions/{rev}/deployments`, two permissions are required: * `apigee.deployments.create` on the resource `organizations/{org}/environments/{env}` * `apigee.sharedflowrevisions.deploy` on the resource `organizations/{org}/sharedflows/{sf}/revisions/{rev}`
Args:
name: string, Required. Name of the shared flow revision to deploy in the following format: `organizations/{org}/environments/{env}/sharedflows/{sharedflow}/revisions/{rev}` (required)
override: boolean, Flag that specifies whether the new deployment replaces other deployed revisions of the shared flow in the environment. Set `override` to `true` to replace other deployed revisions. By default, `override` is `false` and the deployment is rejected if other revisions of the shared flow are deployed in the environment.
+ serviceAccount: string, Google Cloud IAM service account. The service account represents the identity of the deployed proxy, and determines what permissions it has. The format must be `{ACCOUNT_ID}@{PROJECT}.iam.gserviceaccount.com`.
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -167,6 +168,7 @@ Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists.
},
],
+ "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`.
"state": "A String", # Current state of the deployment. This field is not populated in List APIs.
}
@@ -245,6 +247,7 @@ Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists.
},
],
+ "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`.
"state": "A String", # Current state of the deployment. This field is not populated in List APIs.
}
diff --git a/docs/dyn/apigee_v1.organizations.html b/docs/dyn/apigee_v1.organizations.html
index cfbaa2ea9c5..c755cc1951f 100644
--- a/docs/dyn/apigee_v1.organizations.html
+++ b/docs/dyn/apigee_v1.organizations.html
@@ -216,9 +216,6 @@ Method Details
"advancedApiOpsConfig": { # Configuration for the Advanced API Ops add-on. # Configuration for the Advanced API Ops add-on.
"enabled": True or False, # Flag that specifies whether the Advanced API Ops add-on is enabled.
},
- "integrationConfig": { # Configuration for the Integration add-on. # Configuration for the Integration add-on.
- "enabled": True or False, # Flag that specifies whether the Integration add-on is enabled.
- },
"monetizationConfig": { # Configuration for the Monetization add-on. # Configuration for the Monetization add-on.
"enabled": True or False, # Flag that specifies whether the Monetization add-on is enabled.
},
@@ -340,9 +337,6 @@ Method Details
"advancedApiOpsConfig": { # Configuration for the Advanced API Ops add-on. # Configuration for the Advanced API Ops add-on.
"enabled": True or False, # Flag that specifies whether the Advanced API Ops add-on is enabled.
},
- "integrationConfig": { # Configuration for the Integration add-on. # Configuration for the Integration add-on.
- "enabled": True or False, # Flag that specifies whether the Integration add-on is enabled.
- },
"monetizationConfig": { # Configuration for the Monetization add-on. # Configuration for the Monetization add-on.
"enabled": True or False, # Flag that specifies whether the Monetization add-on is enabled.
},
@@ -517,9 +511,6 @@ Method Details
"advancedApiOpsConfig": { # Configuration for the Advanced API Ops add-on. # Configuration for the Advanced API Ops add-on.
"enabled": True or False, # Flag that specifies whether the Advanced API Ops add-on is enabled.
},
- "integrationConfig": { # Configuration for the Integration add-on. # Configuration for the Integration add-on.
- "enabled": True or False, # Flag that specifies whether the Integration add-on is enabled.
- },
"monetizationConfig": { # Configuration for the Monetization add-on. # Configuration for the Monetization add-on.
"enabled": True or False, # Flag that specifies whether the Monetization add-on is enabled.
},
@@ -601,9 +592,6 @@ Method Details
"advancedApiOpsConfig": { # Configuration for the Advanced API Ops add-on. # Configuration for the Advanced API Ops add-on.
"enabled": True or False, # Flag that specifies whether the Advanced API Ops add-on is enabled.
},
- "integrationConfig": { # Configuration for the Integration add-on. # Configuration for the Integration add-on.
- "enabled": True or False, # Flag that specifies whether the Integration add-on is enabled.
- },
"monetizationConfig": { # Configuration for the Monetization add-on. # Configuration for the Monetization add-on.
"enabled": True or False, # Flag that specifies whether the Monetization add-on is enabled.
},
@@ -654,9 +642,6 @@ Method Details
"advancedApiOpsConfig": { # Configuration for the Advanced API Ops add-on. # Configuration for the Advanced API Ops add-on.
"enabled": True or False, # Flag that specifies whether the Advanced API Ops add-on is enabled.
},
- "integrationConfig": { # Configuration for the Integration add-on. # Configuration for the Integration add-on.
- "enabled": True or False, # Flag that specifies whether the Integration add-on is enabled.
- },
"monetizationConfig": { # Configuration for the Monetization add-on. # Configuration for the Monetization add-on.
"enabled": True or False, # Flag that specifies whether the Monetization add-on is enabled.
},
diff --git a/docs/dyn/apigee_v1.organizations.instances.html b/docs/dyn/apigee_v1.organizations.instances.html
index c8b33a55aa3..a423ce20078 100644
--- a/docs/dyn/apigee_v1.organizations.instances.html
+++ b/docs/dyn/apigee_v1.organizations.instances.html
@@ -107,6 +107,9 @@ Instance Methods
list_next(previous_request, previous_response)
Retrieves the next page of results.
+
+ patch(name, body=None, updateMask=None, x__xgafv=None)
+Updates an Apigee runtime instance. You can update the fields described in NodeConfig. No other fields will be updated. **Note:** Not supported for Apigee hybrid.
reportStatus(instance, body=None, x__xgafv=None)
Reports the latest status for a runtime instance.
@@ -282,6 +285,59 @@ Method Details
+patch(name, body=None, updateMask=None, x__xgafv=None)
+ Updates an Apigee runtime instance. You can update the fields described in NodeConfig. No other fields will be updated. **Note:** Not supported for Apigee hybrid.
+
+Args:
+ name: string, Required. Name of the instance. Use the following structure in your request: `organizations/{org}/instances/{instance}`. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Apigee runtime instance.
+ "createdAt": "A String", # Output only. Time the instance was created in milliseconds since epoch.
+ "description": "A String", # Optional. Description of the instance.
+ "diskEncryptionKeyName": "A String", # Customer Managed Encryption Key (CMEK) used for disk and volume encryption. Required for Apigee paid subscriptions only. Use the following format: `projects/([^/]+)/locations/([^/]+)/keyRings/([^/]+)/cryptoKeys/([^/]+)`
+ "displayName": "A String", # Optional. Display name for the instance.
+ "host": "A String", # Output only. Internal hostname or IP address of the Apigee endpoint used by clients to connect to the service.
+ "lastModifiedAt": "A String", # Output only. Time the instance was last modified in milliseconds since epoch.
+ "location": "A String", # Required. Compute Engine location where the instance resides.
+ "name": "A String", # Required. Resource ID of the instance. Values must match the regular expression `^a-z{0,30}[a-z\d]$`.
+ "peeringCidrRange": "A String", # Optional. Size of the CIDR block range that will be reserved by the instance. PAID organizations support `SLASH_16` to `SLASH_20` and defaults to `SLASH_16`. Evaluation organizations support only `SLASH_23`.
+ "port": "A String", # Output only. Port number of the exposed Apigee endpoint.
+ "state": "A String", # Output only. State of the instance. Values other than `ACTIVE` means the resource is not ready to use.
+}
+
+ updateMask: string, List of fields to be updated.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+reportStatus(instance, body=None, x__xgafv=None)
Reports the latest status for a runtime instance. diff --git a/docs/dyn/apigee_v1.organizations.sharedflows.deployments.html b/docs/dyn/apigee_v1.organizations.sharedflows.deployments.html index 145fac017c4..f7fd8406361 100644 --- a/docs/dyn/apigee_v1.organizations.sharedflows.deployments.html +++ b/docs/dyn/apigee_v1.organizations.sharedflows.deployments.html @@ -162,6 +162,7 @@@@ -289,6 +301,9 @@Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists. }, ], + "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`. "state": "A String", # Current state of the deployment. This field is not populated in List APIs. }, ], diff --git a/docs/dyn/apigee_v1.organizations.sharedflows.revisions.deployments.html b/docs/dyn/apigee_v1.organizations.sharedflows.revisions.deployments.html index b97e818f45f..ff63b5841e5 100644 --- a/docs/dyn/apigee_v1.organizations.sharedflows.revisions.deployments.html +++ b/docs/dyn/apigee_v1.organizations.sharedflows.revisions.deployments.html @@ -162,6 +162,7 @@Method Details
"environmentGroup": "A String", # Name of the environment group in which this conflict exists. }, ], + "serviceAccount": "A String", # The full resource name of Cloud IAM Service Account that this deployment is using, eg, `projects/-/serviceAccounts/{email}`. "state": "A String", # Current state of the deployment. This field is not populated in List APIs. }, ], diff --git a/docs/dyn/appengine_v1beta.apps.services.versions.html b/docs/dyn/appengine_v1beta.apps.services.versions.html index 205b0d2b28a..6771a5aec10 100644 --- a/docs/dyn/appengine_v1beta.apps.services.versions.html +++ b/docs/dyn/appengine_v1beta.apps.services.versions.html @@ -290,6 +290,7 @@Method Details
"forwardedPorts": [ # List of ports, or port pairs, to forward from the virtual machine to the application container. Only applicable in the App Engine flexible environment. "A String", ], + "instanceIpMode": "A String", # The IP mode for instances. Only applicable in the App Engine flexible environment. "instanceTag": "A String", # Tag to apply to the instance during creation. Only applicable in the App Engine flexible environment. "name": "A String", # Google Compute Engine network where the virtual machines are created. Specify the short name, not the resource path.Defaults to default. "sessionAffinity": True or False, # Enable session affinity. Only applicable in the App Engine flexible environment. @@ -596,6 +597,7 @@Method Details
"forwardedPorts": [ # List of ports, or port pairs, to forward from the virtual machine to the application container. Only applicable in the App Engine flexible environment. "A String", ], + "instanceIpMode": "A String", # The IP mode for instances. Only applicable in the App Engine flexible environment. "instanceTag": "A String", # Tag to apply to the instance during creation. Only applicable in the App Engine flexible environment. "name": "A String", # Google Compute Engine network where the virtual machines are created. Specify the short name, not the resource path.Defaults to default. "sessionAffinity": True or False, # Enable session affinity. Only applicable in the App Engine flexible environment. @@ -841,6 +843,7 @@Method Details
"forwardedPorts": [ # List of ports, or port pairs, to forward from the virtual machine to the application container. Only applicable in the App Engine flexible environment. "A String", ], + "instanceIpMode": "A String", # The IP mode for instances. Only applicable in the App Engine flexible environment. "instanceTag": "A String", # Tag to apply to the instance during creation. Only applicable in the App Engine flexible environment. "name": "A String", # Google Compute Engine network where the virtual machines are created. Specify the short name, not the resource path.Defaults to default. "sessionAffinity": True or False, # Enable session affinity. Only applicable in the App Engine flexible environment. @@ -1089,6 +1092,7 @@Method Details
"forwardedPorts": [ # List of ports, or port pairs, to forward from the virtual machine to the application container. Only applicable in the App Engine flexible environment. "A String", ], + "instanceIpMode": "A String", # The IP mode for instances. Only applicable in the App Engine flexible environment. "instanceTag": "A String", # Tag to apply to the instance during creation. Only applicable in the App Engine flexible environment. "name": "A String", # Google Compute Engine network where the virtual machines are created. Specify the short name, not the resource path.Defaults to default. "sessionAffinity": True or False, # Enable session affinity. Only applicable in the App Engine flexible environment. diff --git a/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.html b/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.html index 4216667b4ac..794a247992a 100644 --- a/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.html +++ b/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.html @@ -284,7 +284,7 @@Method Details
Args: parent: string, The name of the parent resource whose repositories will be listed. (required) - pageSize: integer, The maximum number of repositories to return. Maximum page size is 10,000. + pageSize: integer, The maximum number of repositories to return. Maximum page size is 1,000. pageToken: string, The next_page_token value returned from a previous list request, if any. x__xgafv: string, V1 error format. Allowed values diff --git a/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.packages.html b/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.packages.html index af1f6832078..a4998549a30 100644 --- a/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.packages.html +++ b/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.packages.html @@ -168,7 +168,7 @@Method Details
Args: parent: string, The name of the parent resource whose packages will be listed. (required) - pageSize: integer, The maximum number of packages to return. Maximum page size is 10,000. + pageSize: integer, The maximum number of packages to return. Maximum page size is 1,000. pageToken: string, The next_page_token value returned from a previous list request, if any. x__xgafv: string, V1 error format. Allowed values diff --git a/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.packages.versions.html b/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.packages.versions.html index d515bd7b0b4..119634d464a 100644 --- a/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.packages.versions.html +++ b/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.packages.versions.html @@ -174,7 +174,7 @@Method Details
Args: parent: string, The name of the parent resource whose versions will be listed. (required) orderBy: string, Optional. Sorting field and order - pageSize: integer, The maximum number of versions to return. Maximum page size is 10,000. + pageSize: integer, The maximum number of versions to return. Maximum page size is 1,000. pageToken: string, The next_page_token value returned from a previous list request, if any. view: string, The view that should be returned in the response. Allowed values diff --git a/docs/dyn/assuredworkloads_v1.organizations.locations.workloads.html b/docs/dyn/assuredworkloads_v1.organizations.locations.workloads.html index cf6360f54d3..0fb6d01608c 100644 --- a/docs/dyn/assuredworkloads_v1.organizations.locations.workloads.html +++ b/docs/dyn/assuredworkloads_v1.organizations.locations.workloads.html @@ -124,7 +124,7 @@Method Details
"a_key": "A String", }, "name": "A String", # Optional. The resource name of the workload. Format: organizations/{organization}/locations/{location}/workloads/{workload} Read-only. - "provisionedResourcesParent": "A String", # Input only. The parent resource for the resources managed by this Assured Workload. May be either an organization or a folder. Must be the same or a child of the Workload parent. If not specified all resources are created under the Workload parent. Formats: folders/{folder_id} organizations/{organization_id} + "provisionedResourcesParent": "A String", # Input only. The parent resource for the resources managed by this Assured Workload. May be either empty or a folder resource which is a child of the Workload parent. If not specified all resources are created under the parent organization. Format: folders/{folder_id} "resourceSettings": [ # Input only. Resource properties that are used to customize workload resources. These properties (such as custom project id) will be used to create workload resources if possible. This field is optional. { # Represent the custom settings for the resources to be created. "displayName": "A String", # User-assigned resource display name. If not empty it will be used to create a resource with the specified name. @@ -217,7 +217,7 @@Method Details
"a_key": "A String", }, "name": "A String", # Optional. The resource name of the workload. Format: organizations/{organization}/locations/{location}/workloads/{workload} Read-only. - "provisionedResourcesParent": "A String", # Input only. The parent resource for the resources managed by this Assured Workload. May be either an organization or a folder. Must be the same or a child of the Workload parent. If not specified all resources are created under the Workload parent. Formats: folders/{folder_id} organizations/{organization_id} + "provisionedResourcesParent": "A String", # Input only. The parent resource for the resources managed by this Assured Workload. May be either empty or a folder resource which is a child of the Workload parent. If not specified all resources are created under the parent organization. Format: folders/{folder_id} "resourceSettings": [ # Input only. Resource properties that are used to customize workload resources. These properties (such as custom project id) will be used to create workload resources if possible. This field is optional. { # Represent the custom settings for the resources to be created. "displayName": "A String", # User-assigned resource display name. If not empty it will be used to create a resource with the specified name. @@ -268,7 +268,7 @@Method Details
"a_key": "A String", }, "name": "A String", # Optional. The resource name of the workload. Format: organizations/{organization}/locations/{location}/workloads/{workload} Read-only. - "provisionedResourcesParent": "A String", # Input only. The parent resource for the resources managed by this Assured Workload. May be either an organization or a folder. Must be the same or a child of the Workload parent. If not specified all resources are created under the Workload parent. Formats: folders/{folder_id} organizations/{organization_id} + "provisionedResourcesParent": "A String", # Input only. The parent resource for the resources managed by this Assured Workload. May be either empty or a folder resource which is a child of the Workload parent. If not specified all resources are created under the parent organization. Format: folders/{folder_id} "resourceSettings": [ # Input only. Resource properties that are used to customize workload resources. These properties (such as custom project id) will be used to create workload resources if possible. This field is optional. { # Represent the custom settings for the resources to be created. "displayName": "A String", # User-assigned resource display name. If not empty it will be used to create a resource with the specified name. @@ -324,7 +324,7 @@Method Details
"a_key": "A String", }, "name": "A String", # Optional. The resource name of the workload. Format: organizations/{organization}/locations/{location}/workloads/{workload} Read-only. - "provisionedResourcesParent": "A String", # Input only. The parent resource for the resources managed by this Assured Workload. May be either an organization or a folder. Must be the same or a child of the Workload parent. If not specified all resources are created under the Workload parent. Formats: folders/{folder_id} organizations/{organization_id} + "provisionedResourcesParent": "A String", # Input only. The parent resource for the resources managed by this Assured Workload. May be either empty or a folder resource which is a child of the Workload parent. If not specified all resources are created under the parent organization. Format: folders/{folder_id} "resourceSettings": [ # Input only. Resource properties that are used to customize workload resources. These properties (such as custom project id) will be used to create workload resources if possible. This field is optional. { # Represent the custom settings for the resources to be created. "displayName": "A String", # User-assigned resource display name. If not empty it will be used to create a resource with the specified name. @@ -363,7 +363,7 @@Method Details
"a_key": "A String", }, "name": "A String", # Optional. The resource name of the workload. Format: organizations/{organization}/locations/{location}/workloads/{workload} Read-only. - "provisionedResourcesParent": "A String", # Input only. The parent resource for the resources managed by this Assured Workload. May be either an organization or a folder. Must be the same or a child of the Workload parent. If not specified all resources are created under the Workload parent. Formats: folders/{folder_id} organizations/{organization_id} + "provisionedResourcesParent": "A String", # Input only. The parent resource for the resources managed by this Assured Workload. May be either empty or a folder resource which is a child of the Workload parent. If not specified all resources are created under the parent organization. Format: folders/{folder_id} "resourceSettings": [ # Input only. Resource properties that are used to customize workload resources. These properties (such as custom project id) will be used to create workload resources if possible. This field is optional. { # Represent the custom settings for the resources to be created. "displayName": "A String", # User-assigned resource display name. If not empty it will be used to create a resource with the specified name. diff --git a/docs/dyn/bigquery_v2.datasets.html b/docs/dyn/bigquery_v2.datasets.html index 40b40207c3e..61e04a26cff 100644 --- a/docs/dyn/bigquery_v2.datasets.html +++ b/docs/dyn/bigquery_v2.datasets.html @@ -172,6 +172,7 @@Method Details
"etag": "A String", # [Output-only] A hash of the resource. "friendlyName": "A String", # [Optional] A descriptive name for the dataset. "id": "A String", # [Output-only] The fully-qualified unique name of the dataset in the format projectId:datasetId. The dataset name without the project name is given in the datasetId field. When creating a new dataset, leave this field blank, and instead specify the datasetId field. + "isCaseInsensitive": True or False, # [Optional] Indicates if table names are case insensitive in the dataset. "kind": "bigquery#dataset", # [Output-only] The resource type. "labels": { # The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See Creating and Updating Dataset Labels for more information. "a_key": "A String", @@ -238,6 +239,7 @@Method Details
"etag": "A String", # [Output-only] A hash of the resource. "friendlyName": "A String", # [Optional] A descriptive name for the dataset. "id": "A String", # [Output-only] The fully-qualified unique name of the dataset in the format projectId:datasetId. The dataset name without the project name is given in the datasetId field. When creating a new dataset, leave this field blank, and instead specify the datasetId field. + "isCaseInsensitive": True or False, # [Optional] Indicates if table names are case insensitive in the dataset. "kind": "bigquery#dataset", # [Output-only] The resource type. "labels": { # The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See Creating and Updating Dataset Labels for more information. "a_key": "A String", @@ -298,6 +300,7 @@Method Details
"etag": "A String", # [Output-only] A hash of the resource. "friendlyName": "A String", # [Optional] A descriptive name for the dataset. "id": "A String", # [Output-only] The fully-qualified unique name of the dataset in the format projectId:datasetId. The dataset name without the project name is given in the datasetId field. When creating a new dataset, leave this field blank, and instead specify the datasetId field. + "isCaseInsensitive": True or False, # [Optional] Indicates if table names are case insensitive in the dataset. "kind": "bigquery#dataset", # [Output-only] The resource type. "labels": { # The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See Creating and Updating Dataset Labels for more information. "a_key": "A String", @@ -415,6 +418,7 @@Method Details
"etag": "A String", # [Output-only] A hash of the resource. "friendlyName": "A String", # [Optional] A descriptive name for the dataset. "id": "A String", # [Output-only] The fully-qualified unique name of the dataset in the format projectId:datasetId. The dataset name without the project name is given in the datasetId field. When creating a new dataset, leave this field blank, and instead specify the datasetId field. + "isCaseInsensitive": True or False, # [Optional] Indicates if table names are case insensitive in the dataset. "kind": "bigquery#dataset", # [Output-only] The resource type. "labels": { # The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See Creating and Updating Dataset Labels for more information. "a_key": "A String", @@ -475,6 +479,7 @@Method Details
"etag": "A String", # [Output-only] A hash of the resource. "friendlyName": "A String", # [Optional] A descriptive name for the dataset. "id": "A String", # [Output-only] The fully-qualified unique name of the dataset in the format projectId:datasetId. The dataset name without the project name is given in the datasetId field. When creating a new dataset, leave this field blank, and instead specify the datasetId field. + "isCaseInsensitive": True or False, # [Optional] Indicates if table names are case insensitive in the dataset. "kind": "bigquery#dataset", # [Output-only] The resource type. "labels": { # The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See Creating and Updating Dataset Labels for more information. "a_key": "A String", @@ -542,6 +547,7 @@Method Details
"etag": "A String", # [Output-only] A hash of the resource. "friendlyName": "A String", # [Optional] A descriptive name for the dataset. "id": "A String", # [Output-only] The fully-qualified unique name of the dataset in the format projectId:datasetId. The dataset name without the project name is given in the datasetId field. When creating a new dataset, leave this field blank, and instead specify the datasetId field. + "isCaseInsensitive": True or False, # [Optional] Indicates if table names are case insensitive in the dataset. "kind": "bigquery#dataset", # [Output-only] The resource type. "labels": { # The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See Creating and Updating Dataset Labels for more information. "a_key": "A String", @@ -602,6 +608,7 @@Method Details
"etag": "A String", # [Output-only] A hash of the resource. "friendlyName": "A String", # [Optional] A descriptive name for the dataset. "id": "A String", # [Output-only] The fully-qualified unique name of the dataset in the format projectId:datasetId. The dataset name without the project name is given in the datasetId field. When creating a new dataset, leave this field blank, and instead specify the datasetId field. + "isCaseInsensitive": True or False, # [Optional] Indicates if table names are case insensitive in the dataset. "kind": "bigquery#dataset", # [Output-only] The resource type. "labels": { # The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See Creating and Updating Dataset Labels for more information. "a_key": "A String", diff --git a/docs/dyn/bigquery_v2.jobs.html b/docs/dyn/bigquery_v2.jobs.html index 345a388ab3a..4c2222a3885 100644 --- a/docs/dyn/bigquery_v2.jobs.html +++ b/docs/dyn/bigquery_v2.jobs.html @@ -236,6 +236,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -375,6 +376,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -404,6 +406,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -596,6 +599,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -860,6 +864,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -999,6 +1004,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -1028,6 +1034,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1220,6 +1227,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1389,6 +1397,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1553,6 +1562,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1692,6 +1702,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -1721,6 +1732,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1913,6 +1925,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -2152,6 +2165,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -2291,6 +2305,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -2320,6 +2335,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -2512,6 +2528,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -2774,6 +2791,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -2913,6 +2931,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -2942,6 +2961,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -3139,6 +3159,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -3381,6 +3402,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema diff --git a/docs/dyn/bigquery_v2.tables.html b/docs/dyn/bigquery_v2.tables.html index 8c4f73d9b05..e19a427cb1b 100644 --- a/docs/dyn/bigquery_v2.tables.html +++ b/docs/dyn/bigquery_v2.tables.html @@ -181,6 +181,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -210,6 +211,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -303,6 +305,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -467,6 +470,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -496,6 +500,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -589,6 +594,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -694,6 +700,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -723,6 +730,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -816,6 +824,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1004,6 +1013,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -1033,6 +1043,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1126,6 +1137,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1231,6 +1243,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -1260,6 +1273,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1353,6 +1367,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1573,6 +1588,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -1602,6 +1618,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1695,6 +1712,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1800,6 +1818,7 @@Method Details
"allowQuotedNewlines": True or False, # [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false. "encoding": "A String", # [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties. "fieldDelimiter": "A String", # [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (','). + "null_marker": "A String", # [Optional] An custom string that will represent a NULL value in CSV import data. "quote": """, # [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true. "skipLeadingRows": "A String", # [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema. }, @@ -1829,6 +1848,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema @@ -1922,6 +1942,7 @@Method Details
"A String", ], }, + "collationSpec": "A String", # Optional. Collation specification of the field. It only can be set on string type field. "description": "A String", # [Optional] The field description. The maximum length is 1,024 characters. "fields": [ # [Optional] Describes the nested schema fields if the type property is set to RECORD. # Object with schema name: TableFieldSchema diff --git a/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.html b/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.html index b5b36034aea..8a2a0aec067 100644 --- a/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.html +++ b/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.html @@ -93,7 +93,7 @@Instance Methods
Returns information about a data transfer config.
-
list(parent, dataSourceIds=None, pageSize=None, pageToken=None, x__xgafv=None)Returns information about all data transfers in the project.
+Returns information about all transfer configs owned by a project in the specified location.
list_next(previous_request, previous_response)Retrieves the next page of results.
@@ -249,7 +249,7 @@Method Details
@@ -254,6 +263,9 @@list(parent, dataSourceIds=None, pageSize=None, pageToken=None, x__xgafv=None)-Returns information about all data transfers in the project. +Returns information about all transfer configs owned by a project in the specified location. Args: parent: string, Required. The BigQuery project id for which data sources should be returned: `projects/{project_id}` or `projects/{project_id}/locations/{location_id}` (required) diff --git a/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.html b/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.html index acff0cc6321..4862f363747 100644 --- a/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.html +++ b/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.html @@ -93,7 +93,7 @@Instance Methods
Returns information about a data transfer config.
-
list(parent, dataSourceIds=None, pageSize=None, pageToken=None, x__xgafv=None)Returns information about all data transfers in the project.
+Returns information about all transfer configs owned by a project in the specified location.
list_next(previous_request, previous_response)Retrieves the next page of results.
@@ -249,7 +249,7 @@Method Details
list(parent, dataSourceIds=None, pageSize=None, pageToken=None, x__xgafv=None)-Returns information about all data transfers in the project. +Returns information about all transfer configs owned by a project in the specified location. Args: parent: string, Required. The BigQuery project id for which data sources should be returned: `projects/{project_id}` or `projects/{project_id}/locations/{location_id}` (required) diff --git a/docs/dyn/chromemanagement_v1.customers.apps.android.html b/docs/dyn/chromemanagement_v1.customers.apps.android.html new file mode 100644 index 00000000000..5a591b90c58 --- /dev/null +++ b/docs/dyn/chromemanagement_v1.customers.apps.android.html @@ -0,0 +1,158 @@ + + + +Chrome Management API . customers . apps . android
+Instance Methods
++
+close()Close httplib2 connections.
+ +Get a specific app for a customer by its resource name.
+Method Details
+++ +close()+Close httplib2 connections.+++ + \ No newline at end of file diff --git a/docs/dyn/chromemanagement_v1.customers.apps.chrome.html b/docs/dyn/chromemanagement_v1.customers.apps.chrome.html new file mode 100644 index 00000000000..acc42c502a8 --- /dev/null +++ b/docs/dyn/chromemanagement_v1.customers.apps.chrome.html @@ -0,0 +1,158 @@ + + + +get(name, x__xgafv=None)+Get a specific app for a customer by its resource name. + +Args: + name: string, Required. The app for which details are being queried. Examples: "customers/my_customer/apps/chrome/gmbmikajjgmnabiglmofipeabaddhgne@2.1.2" for the Save to Google Drive Chrome extension version 2.1.2, "customers/my_customer/apps/android/com.google.android.apps.docs" for the Google Drive Android app's latest version. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Resource representing app details. + "androidAppInfo": { # Android app information. # Output only. Android app information. + "permissions": [ # Output only. Permissions requested by an Android app. + { # Permission requested by an Android app. + "type": "A String", # Output only. The type of the permission. + }, + ], + }, + "appId": "A String", # Output only. Unique store identifier for the item. Examples: "gmbmikajjgmnabiglmofipeabaddhgne" for the Save to Google Drive Chrome extension, "com.google.android.apps.docs" for the Google Drive Android app. + "chromeAppInfo": { # Chrome Web Store app information. # Output only. Chrome Web Store app information. + "googleOwned": True or False, # Output only. Whether the app or extension is built and maintained by Google. Version-specific field that will only be set when the requested app version is found. + "isCwsHosted": True or False, # Output only. Whether the app or extension is in a published state in the Chrome Web Store. + "isTheme": True or False, # Output only. Whether the app or extension is a theme. + "minUserCount": 42, # Output only. The minimum number of users using this app. + "permissions": [ # Output only. Every custom permission requested by the app. Version-specific field that will only be set when the requested app version is found. + { # Permission requested by a Chrome app or extension. + "accessUserData": True or False, # Output only. If available, whether this permissions grants the app/extension access to user data. + "documentationUri": "A String", # Output only. If available, a URI to a page that has documentation for the current permission. + "type": "A String", # Output only. The type of the permission. + }, + ], + "siteAccess": [ # Output only. Every permission giving access to domains or broad host patterns. ( e.g. www.google.com). This includes the matches from content scripts as well as hosts in the permissions node of the manifest. Version-specific field that will only be set when the requested app version is found. + { # Represent one host permission. + "hostMatch": "A String", # Output only. This can contain very specific hosts, or patterns like "*.com" for instance. + }, + ], + "supportEnabled": True or False, # Output only. The app developer has enabled support for their app. Version-specific field that will only be set when the requested app version is found. + }, + "description": "A String", # Output only. App's description. + "detailUri": "A String", # Output only. The uri for the detail page of the item. + "displayName": "A String", # Output only. App's display name. + "firstPublishTime": "A String", # Output only. First published time. + "homepageUri": "A String", # Output only. Home page or Website uri. + "iconUri": "A String", # Output only. A link to an image that can be used as an icon for the product. + "isPaidApp": True or False, # Output only. Indicates if the app has to be paid for OR has paid content. + "latestPublishTime": "A String", # Output only. Latest published time. + "name": "A String", # Output only. Format: name=customers/{customer_id}/apps/{chrome|android|web}/{app_id}@{version} + "privacyPolicyUri": "A String", # Output only. The URI pointing to the privacy policy of the app, if it was provided by the developer. Version-specific field that will only be set when the requested app version is found. + "publisher": "A String", # Output only. The publisher of the item. + "reviewNumber": "A String", # Output only. Number of reviews received. Chrome Web Store review information will always be for the latest version of an app. + "reviewRating": 3.14, # Output only. The rating of the app (on 5 stars). Chrome Web Store review information will always be for the latest version of an app. + "revisionId": "A String", # Output only. App version. A new revision is committed whenever a new version of the app is published. + "serviceError": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Output only. Information about a partial service error if applicable. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "type": "A String", # Output only. App type. +}+Chrome Management API . customers . apps . chrome
+Instance Methods
++
+close()Close httplib2 connections.
+ +Get a specific app for a customer by its resource name.
+Method Details
+++ +close()+Close httplib2 connections.+++ + \ No newline at end of file diff --git a/docs/dyn/chromemanagement_v1.customers.apps.html b/docs/dyn/chromemanagement_v1.customers.apps.html new file mode 100644 index 00000000000..afccf9a0913 --- /dev/null +++ b/docs/dyn/chromemanagement_v1.customers.apps.html @@ -0,0 +1,101 @@ + + + +get(name, x__xgafv=None)+Get a specific app for a customer by its resource name. + +Args: + name: string, Required. The app for which details are being queried. Examples: "customers/my_customer/apps/chrome/gmbmikajjgmnabiglmofipeabaddhgne@2.1.2" for the Save to Google Drive Chrome extension version 2.1.2, "customers/my_customer/apps/android/com.google.android.apps.docs" for the Google Drive Android app's latest version. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Resource representing app details. + "androidAppInfo": { # Android app information. # Output only. Android app information. + "permissions": [ # Output only. Permissions requested by an Android app. + { # Permission requested by an Android app. + "type": "A String", # Output only. The type of the permission. + }, + ], + }, + "appId": "A String", # Output only. Unique store identifier for the item. Examples: "gmbmikajjgmnabiglmofipeabaddhgne" for the Save to Google Drive Chrome extension, "com.google.android.apps.docs" for the Google Drive Android app. + "chromeAppInfo": { # Chrome Web Store app information. # Output only. Chrome Web Store app information. + "googleOwned": True or False, # Output only. Whether the app or extension is built and maintained by Google. Version-specific field that will only be set when the requested app version is found. + "isCwsHosted": True or False, # Output only. Whether the app or extension is in a published state in the Chrome Web Store. + "isTheme": True or False, # Output only. Whether the app or extension is a theme. + "minUserCount": 42, # Output only. The minimum number of users using this app. + "permissions": [ # Output only. Every custom permission requested by the app. Version-specific field that will only be set when the requested app version is found. + { # Permission requested by a Chrome app or extension. + "accessUserData": True or False, # Output only. If available, whether this permissions grants the app/extension access to user data. + "documentationUri": "A String", # Output only. If available, a URI to a page that has documentation for the current permission. + "type": "A String", # Output only. The type of the permission. + }, + ], + "siteAccess": [ # Output only. Every permission giving access to domains or broad host patterns. ( e.g. www.google.com). This includes the matches from content scripts as well as hosts in the permissions node of the manifest. Version-specific field that will only be set when the requested app version is found. + { # Represent one host permission. + "hostMatch": "A String", # Output only. This can contain very specific hosts, or patterns like "*.com" for instance. + }, + ], + "supportEnabled": True or False, # Output only. The app developer has enabled support for their app. Version-specific field that will only be set when the requested app version is found. + }, + "description": "A String", # Output only. App's description. + "detailUri": "A String", # Output only. The uri for the detail page of the item. + "displayName": "A String", # Output only. App's display name. + "firstPublishTime": "A String", # Output only. First published time. + "homepageUri": "A String", # Output only. Home page or Website uri. + "iconUri": "A String", # Output only. A link to an image that can be used as an icon for the product. + "isPaidApp": True or False, # Output only. Indicates if the app has to be paid for OR has paid content. + "latestPublishTime": "A String", # Output only. Latest published time. + "name": "A String", # Output only. Format: name=customers/{customer_id}/apps/{chrome|android|web}/{app_id}@{version} + "privacyPolicyUri": "A String", # Output only. The URI pointing to the privacy policy of the app, if it was provided by the developer. Version-specific field that will only be set when the requested app version is found. + "publisher": "A String", # Output only. The publisher of the item. + "reviewNumber": "A String", # Output only. Number of reviews received. Chrome Web Store review information will always be for the latest version of an app. + "reviewRating": 3.14, # Output only. The rating of the app (on 5 stars). Chrome Web Store review information will always be for the latest version of an app. + "revisionId": "A String", # Output only. App version. A new revision is committed whenever a new version of the app is published. + "serviceError": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Output only. Information about a partial service error if applicable. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "type": "A String", # Output only. App type. +}+Chrome Management API . customers . apps
+Instance Methods
++
+android()+Returns the android Resource.
+ ++
+chrome()+Returns the chrome Resource.
+ ++
+web()+Returns the web Resource.
+ ++
+close()Close httplib2 connections.
+Method Details
+++ + \ No newline at end of file diff --git a/docs/dyn/chromemanagement_v1.customers.apps.web.html b/docs/dyn/chromemanagement_v1.customers.apps.web.html new file mode 100644 index 00000000000..adb6cd125cf --- /dev/null +++ b/docs/dyn/chromemanagement_v1.customers.apps.web.html @@ -0,0 +1,158 @@ + + + +close()+Close httplib2 connections.+Chrome Management API . customers . apps . web
+Instance Methods
++
+close()Close httplib2 connections.
+ +Get a specific app for a customer by its resource name.
+Method Details
+++ +close()+Close httplib2 connections.+++ + \ No newline at end of file diff --git a/docs/dyn/chromemanagement_v1.customers.html b/docs/dyn/chromemanagement_v1.customers.html index 1cddff2ebb1..c32b79f57e9 100644 --- a/docs/dyn/chromemanagement_v1.customers.html +++ b/docs/dyn/chromemanagement_v1.customers.html @@ -74,6 +74,11 @@get(name, x__xgafv=None)+Get a specific app for a customer by its resource name. + +Args: + name: string, Required. The app for which details are being queried. Examples: "customers/my_customer/apps/chrome/gmbmikajjgmnabiglmofipeabaddhgne@2.1.2" for the Save to Google Drive Chrome extension version 2.1.2, "customers/my_customer/apps/android/com.google.android.apps.docs" for the Google Drive Android app's latest version. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Resource representing app details. + "androidAppInfo": { # Android app information. # Output only. Android app information. + "permissions": [ # Output only. Permissions requested by an Android app. + { # Permission requested by an Android app. + "type": "A String", # Output only. The type of the permission. + }, + ], + }, + "appId": "A String", # Output only. Unique store identifier for the item. Examples: "gmbmikajjgmnabiglmofipeabaddhgne" for the Save to Google Drive Chrome extension, "com.google.android.apps.docs" for the Google Drive Android app. + "chromeAppInfo": { # Chrome Web Store app information. # Output only. Chrome Web Store app information. + "googleOwned": True or False, # Output only. Whether the app or extension is built and maintained by Google. Version-specific field that will only be set when the requested app version is found. + "isCwsHosted": True or False, # Output only. Whether the app or extension is in a published state in the Chrome Web Store. + "isTheme": True or False, # Output only. Whether the app or extension is a theme. + "minUserCount": 42, # Output only. The minimum number of users using this app. + "permissions": [ # Output only. Every custom permission requested by the app. Version-specific field that will only be set when the requested app version is found. + { # Permission requested by a Chrome app or extension. + "accessUserData": True or False, # Output only. If available, whether this permissions grants the app/extension access to user data. + "documentationUri": "A String", # Output only. If available, a URI to a page that has documentation for the current permission. + "type": "A String", # Output only. The type of the permission. + }, + ], + "siteAccess": [ # Output only. Every permission giving access to domains or broad host patterns. ( e.g. www.google.com). This includes the matches from content scripts as well as hosts in the permissions node of the manifest. Version-specific field that will only be set when the requested app version is found. + { # Represent one host permission. + "hostMatch": "A String", # Output only. This can contain very specific hosts, or patterns like "*.com" for instance. + }, + ], + "supportEnabled": True or False, # Output only. The app developer has enabled support for their app. Version-specific field that will only be set when the requested app version is found. + }, + "description": "A String", # Output only. App's description. + "detailUri": "A String", # Output only. The uri for the detail page of the item. + "displayName": "A String", # Output only. App's display name. + "firstPublishTime": "A String", # Output only. First published time. + "homepageUri": "A String", # Output only. Home page or Website uri. + "iconUri": "A String", # Output only. A link to an image that can be used as an icon for the product. + "isPaidApp": True or False, # Output only. Indicates if the app has to be paid for OR has paid content. + "latestPublishTime": "A String", # Output only. Latest published time. + "name": "A String", # Output only. Format: name=customers/{customer_id}/apps/{chrome|android|web}/{app_id}@{version} + "privacyPolicyUri": "A String", # Output only. The URI pointing to the privacy policy of the app, if it was provided by the developer. Version-specific field that will only be set when the requested app version is found. + "publisher": "A String", # Output only. The publisher of the item. + "reviewNumber": "A String", # Output only. Number of reviews received. Chrome Web Store review information will always be for the latest version of an app. + "reviewRating": 3.14, # Output only. The rating of the app (on 5 stars). Chrome Web Store review information will always be for the latest version of an app. + "revisionId": "A String", # Output only. App version. A new revision is committed whenever a new version of the app is published. + "serviceError": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Output only. Information about a partial service error if applicable. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "type": "A String", # Output only. App type. +}+Chrome Management API . customers
Instance Methods
++
+apps()+Returns the apps Resource.
+ diff --git a/docs/dyn/chromepolicy_v1.customers.policySchemas.html b/docs/dyn/chromepolicy_v1.customers.policySchemas.html index a7bacc26f7c..01be377d7f4 100644 --- a/docs/dyn/chromepolicy_v1.customers.policySchemas.html +++ b/docs/dyn/chromepolicy_v1.customers.policySchemas.html @@ -173,6 +173,12 @@Method Details
{ # Provides detailed information for a particular field that is part of a PolicySchema. "description": "A String", # Output only. The description for the field. "field": "A String", # Output only. The name of the field for associated with this description. + "fieldDependencies": [ # Output only. Provides a list of fields and the values they must have for this field to be allowed to be set. + { # The field and the value it must have for another field to be allowed to be set. + "sourceField": "A String", # The source field which this field depends on. + "sourceFieldValue": "A String", # The value which the source field must have for this field to be allowed to be set. + }, + ], "inputConstraint": "A String", # Output only. Any input constraints associated on the values for the field. "knownValueDescriptions": [ # Output only. If the field has a set of know values, this field will provide a description for these values. { # Provides detailed information about a known value that is allowed for a particular field in a PolicySchema. @@ -287,6 +293,12 @@Method Details
{ # Provides detailed information for a particular field that is part of a PolicySchema. "description": "A String", # Output only. The description for the field. "field": "A String", # Output only. The name of the field for associated with this description. + "fieldDependencies": [ # Output only. Provides a list of fields and the values they must have for this field to be allowed to be set. + { # The field and the value it must have for another field to be allowed to be set. + "sourceField": "A String", # The source field which this field depends on. + "sourceFieldValue": "A String", # The value which the source field must have for this field to be allowed to be set. + }, + ], "inputConstraint": "A String", # Output only. Any input constraints associated on the values for the field. "knownValueDescriptions": [ # Output only. If the field has a set of know values, this field will provide a description for these values. { # Provides detailed information about a known value that is allowed for a particular field in a PolicySchema. diff --git a/docs/dyn/cloudasset_v1.assets.html b/docs/dyn/cloudasset_v1.assets.html index eaf4d3a3374..784049d6def 100644 --- a/docs/dyn/cloudasset_v1.assets.html +++ b/docs/dyn/cloudasset_v1.assets.html @@ -78,7 +78,7 @@Instance Methods
close()Close httplib2 connections.
+list(parent, assetTypes=None, contentType=None, pageSize=None, pageToken=None, readTime=None, relationshipTypes=None, x__xgafv=None)Lists assets with time and resource types and returns paged results in response.
@@ -90,7 +90,7 @@
list_next(previous_request, previous_response)Method Details
-@@ -215,6 +221,9 @@list(parent, assetTypes=None, contentType=None, pageSize=None, pageToken=None, readTime=None, x__xgafv=None)+list(parent, assetTypes=None, contentType=None, pageSize=None, pageToken=None, readTime=None, relationshipTypes=None, x__xgafv=None)Lists assets with time and resource types and returns paged results in response. Args: @@ -104,9 +104,11 @@Method Details
ORG_POLICY - The Cloud Organization Policy set on an asset. ACCESS_POLICY - The Cloud Access context manager Policy set on an asset. OS_INVENTORY - The runtime OS Inventory information. + RELATIONSHIP - The related resources. pageSize: integer, The maximum number of assets to be returned in a single response. Default is 100, minimum is 1, and maximum is 1000. pageToken: string, The `next_page_token` returned from the previous `ListAssetsResponse`, or unspecified for the first `ListAssetsRequest`. It is a continuation of a prior `ListAssets` call, and the API should return the next page of assets. readTime: string, Timestamp to take an asset snapshot. This can only be set to a timestamp between the current time and the current time minus 35 days (inclusive). If not specified, the current time will be used. Due to delays in resource data collection and indexing, there is a volatile window during which running the same query may get different results. + relationshipTypes: string, A list of relationship types to output, for example: `INSTANCE_TO_INSTANCEGROUP`. This field should only be specified if content_type=RELATIONSHIP. * If specified: it snapshots specified relationships. It returns an error if any of the [relationship_types] doesn't belong to the supported relationship types of the [asset_types] or if any of the [asset_types] doesn't belong to the source types of the [relationship_types]. * Otherwise: it snapshots the supported relationships for all [asset_types] or returns an error if any of the [asset_types] has no relationship support. An unspecified asset types field means all supported asset_types. See [Introduction to Cloud Asset Inventory](https://cloud.google.com/asset-inventory/docs/overview) for all supported asset types and relationship types. (repeated) x__xgafv: string, V1 error format. Allowed values 1 - v1 error format @@ -166,12 +168,12 @@Method Details
}, }, "description": "A String", # Description of the `AccessLevel` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/accessLevels/{short_name}`. The maximum length of the `short_name` component is 50 characters. + "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/accessLevels/{access_level}`. The maximum length of the `access_level` component is 50 characters. "title": "A String", # Human readable title. Must be unique within the Policy. }, "accessPolicy": { # `AccessPolicy` is a container for `AccessLevels` (which define the necessary attributes to use Google Cloud services) and `ServicePerimeters` (which define regions of services able to freely pass data within a perimeter). An access policy is globally visible within an organization, and the restrictions it specifies apply to all projects within an organization. # Please also refer to the [access policy user guide](https://cloud.google.com/access-context-manager/docs/overview#access-policies). "etag": "A String", # Output only. An opaque identifier for the current version of the `AccessPolicy`. This will always be a strongly validated etag, meaning that two Access Polices will be identical if and only if their etags are identical. Clients should not expect this to be in any specific format. - "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{policy_id}` + "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{access_policy}` "parent": "A String", # Required. The parent of this `AccessPolicy` in the Cloud Resource Hierarchy. Currently immutable once created. Format: `organizations/{organization_id}` "title": "A String", # Required. Human readable title. Does not affect behavior. }, @@ -235,7 +237,7 @@Method Details
"version": 42, # Version of the `Policy`. Default version is 0; }, ], - "osInventory": { # The inventory details of a VM. # A representation of runtime OS Inventory information. See [this topic](https://cloud.google.com/compute/docs/instances/os-inventory-management) for more information. + "osInventory": { # This API resource represents the available inventory data for a Compute Engine virtual machine (VM) instance at a given point in time. You can use this API resource to determine the inventory data of your VM. For more information, see [Information provided by OS inventory management](https://cloud.google.com/compute/docs/instances/os-inventory-management#data-collected). # A representation of runtime OS Inventory information. See [this topic](https://cloud.google.com/compute/docs/instances/os-inventory-management) for more information. "items": { # Inventory items related to the VM keyed by an opaque unique identifier for each inventory item. The identifier is unique to each distinct and addressable inventory item and will change, when there is a new package version. "a_key": { # A single piece of inventory on a VM. "availablePackage": { # Software package information of the operating system. # Software package available to be installed on the VM instance. @@ -363,6 +365,7 @@Method Details
"updateTime": "A String", # When this inventory item was last modified. }, }, + "name": "A String", # Output only. The `Inventory` API resource name. Format: `projects/{project_number}/locations/{location}/instances/{instance_id}/inventory` "osInfo": { # Operating system information for the VM. # Base level operating system information for the VM. "architecture": "A String", # The system architecture of the operating system. "hostname": "A String", # The VM hostname. @@ -373,6 +376,24 @@Method Details
"shortName": "A String", # The operating system short name. For example, 'windows' or 'debian'. "version": "A String", # The version of the operating system. }, + "updateTime": "A String", # Output only. Timestamp of the last reported inventory for the VM. + }, + "relatedAssets": { # The detailed related assets with the `relationship_type`. # The related assets of the asset of one relationship type. One asset only represents one type of relationship. + "assets": [ # The peer resources of the relationship. + { # An asset identify in Google Cloud which contains its name, type and ancestors. An asset can be any resource in the Google Cloud [resource hierarchy](https://cloud.google.com/resource-manager/docs/cloud-platform-resource-hierarchy), a resource outside the Google Cloud resource hierarchy (such as Google Kubernetes Engine clusters and objects), or a policy (e.g. Cloud IAM policy). See [Supported asset types](https://cloud.google.com/asset-inventory/docs/supported-asset-types) for more information. + "ancestors": [ # The ancestors of an asset in Google Cloud [resource hierarchy](https://cloud.google.com/resource-manager/docs/cloud-platform-resource-hierarchy), represented as a list of relative resource names. An ancestry path starts with the closest ancestor in the hierarchy and ends at root. Example: `["projects/123456789", "folders/5432", "organizations/1234"]` + "A String", + ], + "asset": "A String", # The full name of the asset. Example: `//compute.googleapis.com/projects/my_project_123/zones/zone1/instances/instance1` See [Resource names](https://cloud.google.com/apis/design/resource_names#full_resource_name) for more information. + "assetType": "A String", # The type of the asset. Example: `compute.googleapis.com/Disk` See [Supported asset types](https://cloud.google.com/asset-inventory/docs/supported-asset-types) for more information. + }, + ], + "relationshipAttributes": { # The relationship attributes which include `type`, `source_resource_type`, `target_resource_type` and `action`. # The detailed relationship attributes. + "action": "A String", # The detail of the relationship, e.g. `contains`, `attaches` + "sourceResourceType": "A String", # The source asset type. Example: `compute.googleapis.com/Instance` + "targetResourceType": "A String", # The target asset type. Example: `compute.googleapis.com/Disk` + "type": "A String", # The unique identifier of the relationship type. Example: `INSTANCE_TO_INSTANCEGROUP` + }, }, "resource": { # A representation of a Google Cloud resource. # A representation of the resource. "data": { # The content of the resource, in which some sensitive fields are removed and may not be present. @@ -387,7 +408,7 @@Method Details
}, "servicePerimeter": { # `ServicePerimeter` describes a set of Google Cloud resources which can freely import and export data amongst themselves, but not export outside of the `ServicePerimeter`. If a request with a source within this `ServicePerimeter` has a target outside of the `ServicePerimeter`, the request will be blocked. Otherwise the request is allowed. There are two types of Service Perimeter - Regular and Bridge. Regular Service Perimeters cannot overlap, a single Google Cloud project can only belong to a single regular Service Perimeter. Service Perimeter Bridges can contain only Google Cloud projects as members, a single Google Cloud project may belong to multiple Service Perimeter Bridges. # Please also refer to the [service perimeter user guide](https://cloud.google.com/vpc-service-controls/docs/overview). "description": "A String", # Description of the `ServicePerimeter` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}` + "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}` "perimeterType": "A String", # Perimeter type indicator. A single project is allowed to be a member of single regular perimeter, but multiple service perimeter bridges. A project cannot be a included in a perimeter bridge without being included in regular perimeter. For perimeter bridges, the restricted service list as well as access level lists must be empty. "spec": { # `ServicePerimeterConfig` specifies a set of Google Cloud resources that describe specific Service Perimeter configuration. # Proposed (or dry run) ServicePerimeter configuration. This configuration allows to specify and test ServicePerimeter configuration without enforcing actual access restrictions. Only allowed to be set when the "use_explicit_dry_run_spec" flag is set. "accessLevels": [ # A list of `AccessLevel` resource names that allow resources within the `ServicePerimeter` to be accessed from the internet. `AccessLevels` listed must be in the same policy as this `ServicePerimeter`. Referencing a nonexistent `AccessLevel` is a syntax error. If no `AccessLevel` names are listed, resources within the perimeter can only be accessed via Google Cloud calls with request origins within the perimeter. Example: `"accessPolicies/MY_POLICY/accessLevels/MY_LEVEL"`. For Service Perimeter Bridge, must be empty. diff --git a/docs/dyn/cloudasset_v1.feeds.html b/docs/dyn/cloudasset_v1.feeds.html index de42d809bfd..11c933267dd 100644 --- a/docs/dyn/cloudasset_v1.feeds.html +++ b/docs/dyn/cloudasset_v1.feeds.html @@ -128,6 +128,9 @@Method Details
}, }, "name": "A String", # Required. The format will be projects/{project_number}/feeds/{client-assigned_feed_identifier} or folders/{folder_number}/feeds/{client-assigned_feed_identifier} or organizations/{organization_number}/feeds/{client-assigned_feed_identifier} The client-assigned feed identifier must be unique within the parent project/folder/organization. + "relationshipTypes": [ # A list of relationship types to output, for example: `INSTANCE_TO_INSTANCEGROUP`. This field should only be specified if content_type=RELATIONSHIP. * If specified: it outputs specified relationship updates on the [asset_names] or the [asset_types]. It returns an error if any of the [relationship_types] doesn't belong to the supported relationship types of the [asset_names] or [asset_types], or any of the [asset_names] or the [asset_types] doesn't belong to the source types of the [relationship_types]. * Otherwise: it outputs the supported relationships of the types of [asset_names] and [asset_types] or returns an error if any of the [asset_names] or the [asset_types] has no replationship support. See [Introduction to Cloud Asset Inventory](https://cloud.google.com/asset-inventory/docs/overview) for all supported asset types and relationship types. + "A String", + ], }, "feedId": "A String", # Required. This is the client-assigned asset feed identifier and it needs to be unique under a specific parent project/folder/organization. } @@ -160,6 +163,9 @@Method Details
}, }, "name": "A String", # Required. The format will be projects/{project_number}/feeds/{client-assigned_feed_identifier} or folders/{folder_number}/feeds/{client-assigned_feed_identifier} or organizations/{organization_number}/feeds/{client-assigned_feed_identifier} The client-assigned feed identifier must be unique within the parent project/folder/organization. + "relationshipTypes": [ # A list of relationship types to output, for example: `INSTANCE_TO_INSTANCEGROUP`. This field should only be specified if content_type=RELATIONSHIP. * If specified: it outputs specified relationship updates on the [asset_names] or the [asset_types]. It returns an error if any of the [relationship_types] doesn't belong to the supported relationship types of the [asset_names] or [asset_types], or any of the [asset_names] or the [asset_types] doesn't belong to the source types of the [relationship_types]. * Otherwise: it outputs the supported relationships of the types of [asset_names] and [asset_types] or returns an error if any of the [asset_names] or the [asset_types] has no replationship support. See [Introduction to Cloud Asset Inventory](https://cloud.google.com/asset-inventory/docs/overview) for all supported asset types and relationship types. + "A String", + ], }Method Details
}, }, "name": "A String", # Required. The format will be projects/{project_number}/feeds/{client-assigned_feed_identifier} or folders/{folder_number}/feeds/{client-assigned_feed_identifier} or organizations/{organization_number}/feeds/{client-assigned_feed_identifier} The client-assigned feed identifier must be unique within the parent project/folder/organization. + "relationshipTypes": [ # A list of relationship types to output, for example: `INSTANCE_TO_INSTANCEGROUP`. This field should only be specified if content_type=RELATIONSHIP. * If specified: it outputs specified relationship updates on the [asset_names] or the [asset_types]. It returns an error if any of the [relationship_types] doesn't belong to the supported relationship types of the [asset_names] or [asset_types], or any of the [asset_names] or the [asset_types] doesn't belong to the source types of the [relationship_types]. * Otherwise: it outputs the supported relationships of the types of [asset_names] and [asset_types] or returns an error if any of the [asset_names] or the [asset_types] has no replationship support. See [Introduction to Cloud Asset Inventory](https://cloud.google.com/asset-inventory/docs/overview) for all supported asset types and relationship types. + "A String", + ], }Method Details
}, }, "name": "A String", # Required. The format will be projects/{project_number}/feeds/{client-assigned_feed_identifier} or folders/{folder_number}/feeds/{client-assigned_feed_identifier} or organizations/{organization_number}/feeds/{client-assigned_feed_identifier} The client-assigned feed identifier must be unique within the parent project/folder/organization. + "relationshipTypes": [ # A list of relationship types to output, for example: `INSTANCE_TO_INSTANCEGROUP`. This field should only be specified if content_type=RELATIONSHIP. * If specified: it outputs specified relationship updates on the [asset_names] or the [asset_types]. It returns an error if any of the [relationship_types] doesn't belong to the supported relationship types of the [asset_names] or [asset_types], or any of the [asset_names] or the [asset_types] doesn't belong to the source types of the [relationship_types]. * Otherwise: it outputs the supported relationships of the types of [asset_names] and [asset_types] or returns an error if any of the [asset_names] or the [asset_types] has no replationship support. See [Introduction to Cloud Asset Inventory](https://cloud.google.com/asset-inventory/docs/overview) for all supported asset types and relationship types. + "A String", + ], }, ], }
analyzeMove(resource, destinationParent=None, view=None, x__xgafv=None)
Analyze moving a resource to a specified destination without kicking off the actual move. The analysis is best effort depending on the user's permissions of viewing different hierarchical policies and configurations. The policies and configuration are subject to change before the actual resource migration takes place.
+batchGetAssetsHistory(parent, assetNames=None, contentType=None, readTimeWindow_endTime=None, readTimeWindow_startTime=None, relationshipTypes=None, x__xgafv=None)
Batch gets the update history of assets that overlap a time window. For IAM_POLICY content, this API outputs history when the asset and its attached IAM POLICY both exist. This can create gaps in the output history. Otherwise, this API outputs history with asset in both non-delete or deleted status. If a specified asset does not exist, this API returns an INVALID_ARGUMENT error.
@@ -477,7 +477,7 @@batchGetAssetsHistory(parent, assetNames=None, contentType=None, readTimeWindow_endTime=None, readTimeWindow_startTime=None, x__xgafv=None)
+ batchGetAssetsHistory(parent, assetNames=None, contentType=None, readTimeWindow_endTime=None, readTimeWindow_startTime=None, relationshipTypes=None, x__xgafv=None)
Batch gets the update history of assets that overlap a time window. For IAM_POLICY content, this API outputs history when the asset and its attached IAM POLICY both exist. This can create gaps in the output history. Otherwise, this API outputs history with asset in both non-delete or deleted status. If a specified asset does not exist, this API returns an INVALID_ARGUMENT error. Args: @@ -491,8 +491,10 @@Method Details
ORG_POLICY - The Cloud Organization Policy set on an asset. ACCESS_POLICY - The Cloud Access context manager Policy set on an asset. OS_INVENTORY - The runtime OS Inventory information. + RELATIONSHIP - The related resources. readTimeWindow_endTime: string, End time of the time window (inclusive). If not specified, the current timestamp is used instead. readTimeWindow_startTime: string, Start time of the time window (exclusive). + relationshipTypes: string, Optional. A list of relationship types to output, for example: `INSTANCE_TO_INSTANCEGROUP`. This field should only be specified if content_type=RELATIONSHIP. * If specified: it outputs specified relationships' history on the [asset_names]. It returns an error if any of the [relationship_types] doesn't belong to the supported relationship types of the [asset_names] or if any of the [asset_names]'s types doesn't belong to the source types of the [relationship_types]. * Otherwise: it outputs the supported relationships' history on the [asset_names] or returns an error if any of the [asset_names]'s types has no relationship support. See [Introduction to Cloud Asset Inventory](https://cloud.google.com/asset-inventory/docs/overview) for all supported asset types and relationship types. (repeated) x__xgafv: string, V1 error format. Allowed values 1 - v1 error format @@ -553,12 +555,12 @@Method Details
}, }, "description": "A String", # Description of the `AccessLevel` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/accessLevels/{short_name}`. The maximum length of the `short_name` component is 50 characters. + "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/accessLevels/{access_level}`. The maximum length of the `access_level` component is 50 characters. "title": "A String", # Human readable title. Must be unique within the Policy. }, "accessPolicy": { # `AccessPolicy` is a container for `AccessLevels` (which define the necessary attributes to use Google Cloud services) and `ServicePerimeters` (which define regions of services able to freely pass data within a perimeter). An access policy is globally visible within an organization, and the restrictions it specifies apply to all projects within an organization. # Please also refer to the [access policy user guide](https://cloud.google.com/access-context-manager/docs/overview#access-policies). "etag": "A String", # Output only. An opaque identifier for the current version of the `AccessPolicy`. This will always be a strongly validated etag, meaning that two Access Polices will be identical if and only if their etags are identical. Clients should not expect this to be in any specific format. - "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{policy_id}` + "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{access_policy}` "parent": "A String", # Required. The parent of this `AccessPolicy` in the Cloud Resource Hierarchy. Currently immutable once created. Format: `organizations/{organization_id}` "title": "A String", # Required. Human readable title. Does not affect behavior. }, @@ -622,7 +624,7 @@Method Details
"version": 42, # Version of the `Policy`. Default version is 0; }, ], - "osInventory": { # The inventory details of a VM. # A representation of runtime OS Inventory information. See [this topic](https://cloud.google.com/compute/docs/instances/os-inventory-management) for more information. + "osInventory": { # This API resource represents the available inventory data for a Compute Engine virtual machine (VM) instance at a given point in time. You can use this API resource to determine the inventory data of your VM. For more information, see [Information provided by OS inventory management](https://cloud.google.com/compute/docs/instances/os-inventory-management#data-collected). # A representation of runtime OS Inventory information. See [this topic](https://cloud.google.com/compute/docs/instances/os-inventory-management) for more information. "items": { # Inventory items related to the VM keyed by an opaque unique identifier for each inventory item. The identifier is unique to each distinct and addressable inventory item and will change, when there is a new package version. "a_key": { # A single piece of inventory on a VM. "availablePackage": { # Software package information of the operating system. # Software package available to be installed on the VM instance. @@ -750,6 +752,7 @@Method Details
"updateTime": "A String", # When this inventory item was last modified. }, }, + "name": "A String", # Output only. The `Inventory` API resource name. Format: `projects/{project_number}/locations/{location}/instances/{instance_id}/inventory` "osInfo": { # Operating system information for the VM. # Base level operating system information for the VM. "architecture": "A String", # The system architecture of the operating system. "hostname": "A String", # The VM hostname. @@ -760,6 +763,24 @@Method Details
"shortName": "A String", # The operating system short name. For example, 'windows' or 'debian'. "version": "A String", # The version of the operating system. }, + "updateTime": "A String", # Output only. Timestamp of the last reported inventory for the VM. + }, + "relatedAssets": { # The detailed related assets with the `relationship_type`. # The related assets of the asset of one relationship type. One asset only represents one type of relationship. + "assets": [ # The peer resources of the relationship. + { # An asset identify in Google Cloud which contains its name, type and ancestors. An asset can be any resource in the Google Cloud [resource hierarchy](https://cloud.google.com/resource-manager/docs/cloud-platform-resource-hierarchy), a resource outside the Google Cloud resource hierarchy (such as Google Kubernetes Engine clusters and objects), or a policy (e.g. Cloud IAM policy). See [Supported asset types](https://cloud.google.com/asset-inventory/docs/supported-asset-types) for more information. + "ancestors": [ # The ancestors of an asset in Google Cloud [resource hierarchy](https://cloud.google.com/resource-manager/docs/cloud-platform-resource-hierarchy), represented as a list of relative resource names. An ancestry path starts with the closest ancestor in the hierarchy and ends at root. Example: `["projects/123456789", "folders/5432", "organizations/1234"]` + "A String", + ], + "asset": "A String", # The full name of the asset. Example: `//compute.googleapis.com/projects/my_project_123/zones/zone1/instances/instance1` See [Resource names](https://cloud.google.com/apis/design/resource_names#full_resource_name) for more information. + "assetType": "A String", # The type of the asset. Example: `compute.googleapis.com/Disk` See [Supported asset types](https://cloud.google.com/asset-inventory/docs/supported-asset-types) for more information. + }, + ], + "relationshipAttributes": { # The relationship attributes which include `type`, `source_resource_type`, `target_resource_type` and `action`. # The detailed relationship attributes. + "action": "A String", # The detail of the relationship, e.g. `contains`, `attaches` + "sourceResourceType": "A String", # The source asset type. Example: `compute.googleapis.com/Instance` + "targetResourceType": "A String", # The target asset type. Example: `compute.googleapis.com/Disk` + "type": "A String", # The unique identifier of the relationship type. Example: `INSTANCE_TO_INSTANCEGROUP` + }, }, "resource": { # A representation of a Google Cloud resource. # A representation of the resource. "data": { # The content of the resource, in which some sensitive fields are removed and may not be present. @@ -774,7 +795,7 @@Method Details
}, "servicePerimeter": { # `ServicePerimeter` describes a set of Google Cloud resources which can freely import and export data amongst themselves, but not export outside of the `ServicePerimeter`. If a request with a source within this `ServicePerimeter` has a target outside of the `ServicePerimeter`, the request will be blocked. Otherwise the request is allowed. There are two types of Service Perimeter - Regular and Bridge. Regular Service Perimeters cannot overlap, a single Google Cloud project can only belong to a single regular Service Perimeter. Service Perimeter Bridges can contain only Google Cloud projects as members, a single Google Cloud project may belong to multiple Service Perimeter Bridges. # Please also refer to the [service perimeter user guide](https://cloud.google.com/vpc-service-controls/docs/overview). "description": "A String", # Description of the `ServicePerimeter` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}` + "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}` "perimeterType": "A String", # Perimeter type indicator. A single project is allowed to be a member of single regular perimeter, but multiple service perimeter bridges. A project cannot be a included in a perimeter bridge without being included in regular perimeter. For perimeter bridges, the restricted service list as well as access level lists must be empty. "spec": { # `ServicePerimeterConfig` specifies a set of Google Cloud resources that describe specific Service Perimeter configuration. # Proposed (or dry run) ServicePerimeter configuration. This configuration allows to specify and test ServicePerimeter configuration without enforcing actual access restrictions. Only allowed to be set when the "use_explicit_dry_run_spec" flag is set. "accessLevels": [ # A list of `AccessLevel` resource names that allow resources within the `ServicePerimeter` to be accessed from the internet. `AccessLevels` listed must be in the same policy as this `ServicePerimeter`. Referencing a nonexistent `AccessLevel` is a syntax error. If no `AccessLevel` names are listed, resources within the perimeter can only be accessed via Google Cloud calls with request origins within the perimeter. Example: `"accessPolicies/MY_POLICY/accessLevels/MY_LEVEL"`. For Service Perimeter Bridge, must be empty. @@ -981,12 +1002,12 @@Method Details
}, }, "description": "A String", # Description of the `AccessLevel` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/accessLevels/{short_name}`. The maximum length of the `short_name` component is 50 characters. + "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/accessLevels/{access_level}`. The maximum length of the `access_level` component is 50 characters. "title": "A String", # Human readable title. Must be unique within the Policy. }, "accessPolicy": { # `AccessPolicy` is a container for `AccessLevels` (which define the necessary attributes to use Google Cloud services) and `ServicePerimeters` (which define regions of services able to freely pass data within a perimeter). An access policy is globally visible within an organization, and the restrictions it specifies apply to all projects within an organization. # Please also refer to the [access policy user guide](https://cloud.google.com/access-context-manager/docs/overview#access-policies). "etag": "A String", # Output only. An opaque identifier for the current version of the `AccessPolicy`. This will always be a strongly validated etag, meaning that two Access Polices will be identical if and only if their etags are identical. Clients should not expect this to be in any specific format. - "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{policy_id}` + "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{access_policy}` "parent": "A String", # Required. The parent of this `AccessPolicy` in the Cloud Resource Hierarchy. Currently immutable once created. Format: `organizations/{organization_id}` "title": "A String", # Required. Human readable title. Does not affect behavior. }, @@ -1050,7 +1071,7 @@Method Details
"version": 42, # Version of the `Policy`. Default version is 0; }, ], - "osInventory": { # The inventory details of a VM. # A representation of runtime OS Inventory information. See [this topic](https://cloud.google.com/compute/docs/instances/os-inventory-management) for more information. + "osInventory": { # This API resource represents the available inventory data for a Compute Engine virtual machine (VM) instance at a given point in time. You can use this API resource to determine the inventory data of your VM. For more information, see [Information provided by OS inventory management](https://cloud.google.com/compute/docs/instances/os-inventory-management#data-collected). # A representation of runtime OS Inventory information. See [this topic](https://cloud.google.com/compute/docs/instances/os-inventory-management) for more information. "items": { # Inventory items related to the VM keyed by an opaque unique identifier for each inventory item. The identifier is unique to each distinct and addressable inventory item and will change, when there is a new package version. "a_key": { # A single piece of inventory on a VM. "availablePackage": { # Software package information of the operating system. # Software package available to be installed on the VM instance. @@ -1178,6 +1199,7 @@Method Details
"updateTime": "A String", # When this inventory item was last modified. }, }, + "name": "A String", # Output only. The `Inventory` API resource name. Format: `projects/{project_number}/locations/{location}/instances/{instance_id}/inventory` "osInfo": { # Operating system information for the VM. # Base level operating system information for the VM. "architecture": "A String", # The system architecture of the operating system. "hostname": "A String", # The VM hostname. @@ -1188,6 +1210,24 @@Method Details
"shortName": "A String", # The operating system short name. For example, 'windows' or 'debian'. "version": "A String", # The version of the operating system. }, + "updateTime": "A String", # Output only. Timestamp of the last reported inventory for the VM. + }, + "relatedAssets": { # The detailed related assets with the `relationship_type`. # The related assets of the asset of one relationship type. One asset only represents one type of relationship. + "assets": [ # The peer resources of the relationship. + { # An asset identify in Google Cloud which contains its name, type and ancestors. An asset can be any resource in the Google Cloud [resource hierarchy](https://cloud.google.com/resource-manager/docs/cloud-platform-resource-hierarchy), a resource outside the Google Cloud resource hierarchy (such as Google Kubernetes Engine clusters and objects), or a policy (e.g. Cloud IAM policy). See [Supported asset types](https://cloud.google.com/asset-inventory/docs/supported-asset-types) for more information. + "ancestors": [ # The ancestors of an asset in Google Cloud [resource hierarchy](https://cloud.google.com/resource-manager/docs/cloud-platform-resource-hierarchy), represented as a list of relative resource names. An ancestry path starts with the closest ancestor in the hierarchy and ends at root. Example: `["projects/123456789", "folders/5432", "organizations/1234"]` + "A String", + ], + "asset": "A String", # The full name of the asset. Example: `//compute.googleapis.com/projects/my_project_123/zones/zone1/instances/instance1` See [Resource names](https://cloud.google.com/apis/design/resource_names#full_resource_name) for more information. + "assetType": "A String", # The type of the asset. Example: `compute.googleapis.com/Disk` See [Supported asset types](https://cloud.google.com/asset-inventory/docs/supported-asset-types) for more information. + }, + ], + "relationshipAttributes": { # The relationship attributes which include `type`, `source_resource_type`, `target_resource_type` and `action`. # The detailed relationship attributes. + "action": "A String", # The detail of the relationship, e.g. `contains`, `attaches` + "sourceResourceType": "A String", # The source asset type. Example: `compute.googleapis.com/Instance` + "targetResourceType": "A String", # The target asset type. Example: `compute.googleapis.com/Disk` + "type": "A String", # The unique identifier of the relationship type. Example: `INSTANCE_TO_INSTANCEGROUP` + }, }, "resource": { # A representation of a Google Cloud resource. # A representation of the resource. "data": { # The content of the resource, in which some sensitive fields are removed and may not be present. @@ -1202,7 +1242,7 @@Method Details
}, "servicePerimeter": { # `ServicePerimeter` describes a set of Google Cloud resources which can freely import and export data amongst themselves, but not export outside of the `ServicePerimeter`. If a request with a source within this `ServicePerimeter` has a target outside of the `ServicePerimeter`, the request will be blocked. Otherwise the request is allowed. There are two types of Service Perimeter - Regular and Bridge. Regular Service Perimeters cannot overlap, a single Google Cloud project can only belong to a single regular Service Perimeter. Service Perimeter Bridges can contain only Google Cloud projects as members, a single Google Cloud project may belong to multiple Service Perimeter Bridges. # Please also refer to the [service perimeter user guide](https://cloud.google.com/vpc-service-controls/docs/overview). "description": "A String", # Description of the `ServicePerimeter` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}` + "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}` "perimeterType": "A String", # Perimeter type indicator. A single project is allowed to be a member of single regular perimeter, but multiple service perimeter bridges. A project cannot be a included in a perimeter bridge without being included in regular perimeter. For perimeter bridges, the restricted service list as well as access level lists must be empty. "spec": { # `ServicePerimeterConfig` specifies a set of Google Cloud resources that describe specific Service Perimeter configuration. # Proposed (or dry run) ServicePerimeter configuration. This configuration allows to specify and test ServicePerimeter configuration without enforcing actual access restrictions. Only allowed to be set when the "use_explicit_dry_run_spec" flag is set. "accessLevels": [ # A list of `AccessLevel` resource names that allow resources within the `ServicePerimeter` to be accessed from the internet. `AccessLevels` listed must be in the same policy as this `ServicePerimeter`. Referencing a nonexistent `AccessLevel` is a syntax error. If no `AccessLevel` names are listed, resources within the perimeter can only be accessed via Google Cloud calls with request origins within the perimeter. Example: `"accessPolicies/MY_POLICY/accessLevels/MY_LEVEL"`. For Service Perimeter Bridge, must be empty. @@ -1404,6 +1444,9 @@Method Details
}, }, "readTime": "A String", # Timestamp to take an asset snapshot. This can only be set to a timestamp between the current time and the current time minus 35 days (inclusive). If not specified, the current time will be used. Due to delays in resource data collection and indexing, there is a volatile window during which running the same query may get different results. + "relationshipTypes": [ # A list of relationship types to export, for example: `INSTANCE_TO_INSTANCEGROUP`. This field should only be specified if content_type=RELATIONSHIP. * If specified: it snapshots specified relationships. It returns an error if any of the [relationship_types] doesn't belong to the supported relationship types of the [asset_types] or if any of the [asset_types] doesn't belong to the source types of the [relationship_types]. * Otherwise: it snapshots the supported relationships for all [asset_types] or returns an error if any of the [asset_types] has no relationship support. An unspecified asset types field means all supported asset_types. See [Introduction to Cloud Asset Inventory](https://cloud.google.com/asset-inventory/docs/overview) for all supported asset types and relationship types. + "A String", + ], } x__xgafv: string, V1 error format. diff --git a/docs/dyn/cloudasset_v1beta1.organizations.html b/docs/dyn/cloudasset_v1beta1.organizations.html index 552d850542c..bd02138ceec 100644 --- a/docs/dyn/cloudasset_v1beta1.organizations.html +++ b/docs/dyn/cloudasset_v1beta1.organizations.html @@ -163,12 +163,12 @@Method Details
}, }, "description": "A String", # Description of the `AccessLevel` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/accessLevels/{short_name}`. The maximum length of the `short_name` component is 50 characters. + "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/accessLevels/{access_level}`. The maximum length of the `access_level` component is 50 characters. "title": "A String", # Human readable title. Must be unique within the Policy. }, "accessPolicy": { # `AccessPolicy` is a container for `AccessLevels` (which define the necessary attributes to use Google Cloud services) and `ServicePerimeters` (which define regions of services able to freely pass data within a perimeter). An access policy is globally visible within an organization, and the restrictions it specifies apply to all projects within an organization. # Please also refer to the [access policy user guide](https://cloud.google.com/access-context-manager/docs/overview#access-policies). "etag": "A String", # Output only. An opaque identifier for the current version of the `AccessPolicy`. This will always be a strongly validated etag, meaning that two Access Polices will be identical if and only if their etags are identical. Clients should not expect this to be in any specific format. - "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{policy_id}` + "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{access_policy}` "parent": "A String", # Required. The parent of this `AccessPolicy` in the Cloud Resource Hierarchy. Currently immutable once created. Format: `organizations/{organization_id}` "title": "A String", # Required. Human readable title. Does not affect behavior. }, @@ -241,7 +241,7 @@Method Details
}, "servicePerimeter": { # `ServicePerimeter` describes a set of Google Cloud resources which can freely import and export data amongst themselves, but not export outside of the `ServicePerimeter`. If a request with a source within this `ServicePerimeter` has a target outside of the `ServicePerimeter`, the request will be blocked. Otherwise the request is allowed. There are two types of Service Perimeter - Regular and Bridge. Regular Service Perimeters cannot overlap, a single Google Cloud project can only belong to a single regular Service Perimeter. Service Perimeter Bridges can contain only Google Cloud projects as members, a single Google Cloud project may belong to multiple Service Perimeter Bridges. # Please also refer to the [service perimeter user guide](https://cloud.google.com/vpc-service-controls/docs/overview). "description": "A String", # Description of the `ServicePerimeter` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}` + "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}` "perimeterType": "A String", # Perimeter type indicator. A single project is allowed to be a member of single regular perimeter, but multiple service perimeter bridges. A project cannot be a included in a perimeter bridge without being included in regular perimeter. For perimeter bridges, the restricted service list as well as access level lists must be empty. "spec": { # `ServicePerimeterConfig` specifies a set of Google Cloud resources that describe specific Service Perimeter configuration. # Proposed (or dry run) ServicePerimeter configuration. This configuration allows to specify and test ServicePerimeter configuration without enforcing actual access restrictions. Only allowed to be set when the "use_explicit_dry_run_spec" flag is set. "accessLevels": [ # A list of `AccessLevel` resource names that allow resources within the `ServicePerimeter` to be accessed from the internet. `AccessLevels` listed must be in the same policy as this `ServicePerimeter`. Referencing a nonexistent `AccessLevel` is a syntax error. If no `AccessLevel` names are listed, resources within the perimeter can only be accessed via Google Cloud calls with request origins within the perimeter. Example: `"accessPolicies/MY_POLICY/accessLevels/MY_LEVEL"`. For Service Perimeter Bridge, must be empty. diff --git a/docs/dyn/cloudasset_v1beta1.projects.html b/docs/dyn/cloudasset_v1beta1.projects.html index 0b85da0ed72..a3c72db1725 100644 --- a/docs/dyn/cloudasset_v1beta1.projects.html +++ b/docs/dyn/cloudasset_v1beta1.projects.html @@ -163,12 +163,12 @@Method Details
}, }, "description": "A String", # Description of the `AccessLevel` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/accessLevels/{short_name}`. The maximum length of the `short_name` component is 50 characters. + "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/accessLevels/{access_level}`. The maximum length of the `access_level` component is 50 characters. "title": "A String", # Human readable title. Must be unique within the Policy. }, "accessPolicy": { # `AccessPolicy` is a container for `AccessLevels` (which define the necessary attributes to use Google Cloud services) and `ServicePerimeters` (which define regions of services able to freely pass data within a perimeter). An access policy is globally visible within an organization, and the restrictions it specifies apply to all projects within an organization. # Please also refer to the [access policy user guide](https://cloud.google.com/access-context-manager/docs/overview#access-policies). "etag": "A String", # Output only. An opaque identifier for the current version of the `AccessPolicy`. This will always be a strongly validated etag, meaning that two Access Polices will be identical if and only if their etags are identical. Clients should not expect this to be in any specific format. - "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{policy_id}` + "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{access_policy}` "parent": "A String", # Required. The parent of this `AccessPolicy` in the Cloud Resource Hierarchy. Currently immutable once created. Format: `organizations/{organization_id}` "title": "A String", # Required. Human readable title. Does not affect behavior. }, @@ -241,7 +241,7 @@Method Details
}, "servicePerimeter": { # `ServicePerimeter` describes a set of Google Cloud resources which can freely import and export data amongst themselves, but not export outside of the `ServicePerimeter`. If a request with a source within this `ServicePerimeter` has a target outside of the `ServicePerimeter`, the request will be blocked. Otherwise the request is allowed. There are two types of Service Perimeter - Regular and Bridge. Regular Service Perimeters cannot overlap, a single Google Cloud project can only belong to a single regular Service Perimeter. Service Perimeter Bridges can contain only Google Cloud projects as members, a single Google Cloud project may belong to multiple Service Perimeter Bridges. # Please also refer to the [service perimeter user guide](https://cloud.google.com/vpc-service-controls/docs/overview). "description": "A String", # Description of the `ServicePerimeter` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}` + "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}` "perimeterType": "A String", # Perimeter type indicator. A single project is allowed to be a member of single regular perimeter, but multiple service perimeter bridges. A project cannot be a included in a perimeter bridge without being included in regular perimeter. For perimeter bridges, the restricted service list as well as access level lists must be empty. "spec": { # `ServicePerimeterConfig` specifies a set of Google Cloud resources that describe specific Service Perimeter configuration. # Proposed (or dry run) ServicePerimeter configuration. This configuration allows to specify and test ServicePerimeter configuration without enforcing actual access restrictions. Only allowed to be set when the "use_explicit_dry_run_spec" flag is set. "accessLevels": [ # A list of `AccessLevel` resource names that allow resources within the `ServicePerimeter` to be accessed from the internet. `AccessLevels` listed must be in the same policy as this `ServicePerimeter`. Referencing a nonexistent `AccessLevel` is a syntax error. If no `AccessLevel` names are listed, resources within the perimeter can only be accessed via Google Cloud calls with request origins within the perimeter. Example: `"accessPolicies/MY_POLICY/accessLevels/MY_LEVEL"`. For Service Perimeter Bridge, must be empty. diff --git a/docs/dyn/cloudasset_v1p5beta1.assets.html b/docs/dyn/cloudasset_v1p5beta1.assets.html index a2c71232183..b3a65e6b0c8 100644 --- a/docs/dyn/cloudasset_v1p5beta1.assets.html +++ b/docs/dyn/cloudasset_v1p5beta1.assets.html @@ -165,12 +165,12 @@Method Details
}, }, "description": "A String", # Description of the `AccessLevel` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/accessLevels/{short_name}`. The maximum length of the `short_name` component is 50 characters. + "name": "A String", # Required. Resource name for the Access Level. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/accessLevels/{access_level}`. The maximum length of the `access_level` component is 50 characters. "title": "A String", # Human readable title. Must be unique within the Policy. }, "accessPolicy": { # `AccessPolicy` is a container for `AccessLevels` (which define the necessary attributes to use Google Cloud services) and `ServicePerimeters` (which define regions of services able to freely pass data within a perimeter). An access policy is globally visible within an organization, and the restrictions it specifies apply to all projects within an organization. # Please also refer to the [access policy user guide](https://cloud.google.com/access-context-manager/docs/overview#access-policies). "etag": "A String", # Output only. An opaque identifier for the current version of the `AccessPolicy`. This will always be a strongly validated etag, meaning that two Access Polices will be identical if and only if their etags are identical. Clients should not expect this to be in any specific format. - "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{policy_id}` + "name": "A String", # Output only. Resource name of the `AccessPolicy`. Format: `accessPolicies/{access_policy}` "parent": "A String", # Required. The parent of this `AccessPolicy` in the Cloud Resource Hierarchy. Currently immutable once created. Format: `organizations/{organization_id}` "title": "A String", # Required. Human readable title. Does not affect behavior. }, @@ -246,7 +246,7 @@Method Details
}, "servicePerimeter": { # `ServicePerimeter` describes a set of Google Cloud resources which can freely import and export data amongst themselves, but not export outside of the `ServicePerimeter`. If a request with a source within this `ServicePerimeter` has a target outside of the `ServicePerimeter`, the request will be blocked. Otherwise the request is allowed. There are two types of Service Perimeter - Regular and Bridge. Regular Service Perimeters cannot overlap, a single Google Cloud project can only belong to a single regular Service Perimeter. Service Perimeter Bridges can contain only Google Cloud projects as members, a single Google Cloud project may belong to multiple Service Perimeter Bridges. # Please also refer to the [service perimeter user guide](https://cloud.google.com/vpc-service-controls/docs/overview). "description": "A String", # Description of the `ServicePerimeter` and its use. Does not affect behavior. - "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{policy_id}/servicePerimeters/{short_name}` + "name": "A String", # Required. Resource name for the ServicePerimeter. The `short_name` component must begin with a letter and only include alphanumeric and '_'. Format: `accessPolicies/{access_policy}/servicePerimeters/{service_perimeter}` "perimeterType": "A String", # Perimeter type indicator. A single project is allowed to be a member of single regular perimeter, but multiple service perimeter bridges. A project cannot be a included in a perimeter bridge without being included in regular perimeter. For perimeter bridges, the restricted service list as well as access level lists must be empty. "spec": { # `ServicePerimeterConfig` specifies a set of Google Cloud resources that describe specific Service Perimeter configuration. # Proposed (or dry run) ServicePerimeter configuration. This configuration allows to specify and test ServicePerimeter configuration without enforcing actual access restrictions. Only allowed to be set when the "use_explicit_dry_run_spec" flag is set. "accessLevels": [ # A list of `AccessLevel` resource names that allow resources within the `ServicePerimeter` to be accessed from the internet. `AccessLevels` listed must be in the same policy as this `ServicePerimeter`. Referencing a nonexistent `AccessLevel` is a syntax error. If no `AccessLevel` names are listed, resources within the perimeter can only be accessed via Google Cloud calls with request origins within the perimeter. Example: `"accessPolicies/MY_POLICY/accessLevels/MY_LEVEL"`. For Service Perimeter Bridge, must be empty. diff --git a/docs/dyn/cloudbuild_v1.html b/docs/dyn/cloudbuild_v1.html index 5a689ed80e2..ed9c61f05c8 100644 --- a/docs/dyn/cloudbuild_v1.html +++ b/docs/dyn/cloudbuild_v1.html @@ -84,6 +84,11 @@Instance Methods
Returns the projects Resource.
++
+v1()+Returns the v1 Resource.
+Close httplib2 connections.
diff --git a/docs/dyn/cloudbuild_v1.projects.builds.html b/docs/dyn/cloudbuild_v1.projects.builds.html index ffa4e8037ae..bf6e40b85b7 100644 --- a/docs/dyn/cloudbuild_v1.projects.builds.html +++ b/docs/dyn/cloudbuild_v1.projects.builds.html @@ -333,7 +333,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -575,7 +575,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -848,7 +848,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -1096,7 +1096,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. diff --git a/docs/dyn/cloudbuild_v1.projects.githubEnterpriseConfigs.html b/docs/dyn/cloudbuild_v1.projects.githubEnterpriseConfigs.html new file mode 100644 index 00000000000..944b2b4d4a0 --- /dev/null +++ b/docs/dyn/cloudbuild_v1.projects.githubEnterpriseConfigs.html @@ -0,0 +1,337 @@ + + + +Cloud Build API . projects . githubEnterpriseConfigs
+Instance Methods
++
+close()Close httplib2 connections.
++
+create(parent, body=None, projectId=None, x__xgafv=None)Create an association between a GCP project and a GitHub Enterprise server. This API is experimental.
++
+delete(name, configId=None, projectId=None, x__xgafv=None)Delete an association between a GCP project and a GitHub Enterprise server. This API is experimental.
++
+get(name, configId=None, projectId=None, x__xgafv=None)Retrieve a GitHubEnterpriseConfig. This API is experimental.
++
+list(parent, projectId=None, x__xgafv=None)List all GitHubEnterpriseConfigs for a given project. This API is experimental.
++
+patch(name, body=None, updateMask=None, x__xgafv=None)Update an association between a GCP project and a GitHub Enterprise server. This API is experimental.
+Method Details
+++ +close()+Close httplib2 connections.+++ +create(parent, body=None, projectId=None, x__xgafv=None)+Create an association between a GCP project and a GitHub Enterprise server. This API is experimental. + +Args: + parent: string, Name of the parent project. For example: projects/{$project_number} or projects/{$project_id} (required) + body: object, The request body. + The object takes the form of: + +{ # GitHubEnterpriseConfig represents a configuration for a GitHub Enterprise server. + "appId": "A String", # Required. The GitHub app id of the Cloud Build app on the GitHub Enterprise server. + "createTime": "A String", # Output only. Time when the installation was associated with the project. + "displayName": "A String", # Name to display for this config. + "hostUrl": "A String", # The URL of the github enterprise host the configuration is for. + "name": "A String", # Optional. The full resource name for the GitHubEnterpriseConfig For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" + "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the GitHub Enterprise server. The VPC network must be enabled for private service connection. This should be set if the GitHub Enterprise server is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the GitHub Enterprise server will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project. + "secrets": { # GitHubEnterpriseSecrets represents the names of all necessary secrets in Secret Manager for a GitHub Enterprise server. Format is: projects//secrets/. # Names of secrets in Secret Manager. + "oauthClientIdName": "A String", # The resource name for the OAuth client ID secret in Secret Manager. + "oauthClientIdVersionName": "A String", # The resource name for the OAuth client ID secret version in Secret Manager. + "oauthSecretName": "A String", # The resource name for the OAuth secret in Secret Manager. + "oauthSecretVersionName": "A String", # The resource name for the OAuth secret secret version in Secret Manager. + "privateKeyName": "A String", # The resource name for the private key secret. + "privateKeyVersionName": "A String", # The resource name for the private key secret version. + "webhookSecretName": "A String", # The resource name for the webhook secret in Secret Manager. + "webhookSecretVersionName": "A String", # The resource name for the webhook secret secret version in Secret Manager. + }, + "sslCa": "A String", # Optional. SSL certificate to use for requests to GitHub Enterprise. + "webhookKey": "A String", # The key that should be attached to webhook calls to the ReceiveWebhook endpoint. +} + + projectId: string, ID of the project. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +delete(name, configId=None, projectId=None, x__xgafv=None)+Delete an association between a GCP project and a GitHub Enterprise server. This API is experimental. + +Args: + name: string, This field should contain the name of the enterprise config resource. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" (required) + configId: string, Unique identifier of the `GitHubEnterpriseConfig` + projectId: string, ID of the project + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +get(name, configId=None, projectId=None, x__xgafv=None)+Retrieve a GitHubEnterpriseConfig. This API is experimental. + +Args: + name: string, This field should contain the name of the enterprise config resource. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" (required) + configId: string, Unique identifier of the `GitHubEnterpriseConfig` + projectId: string, ID of the project + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # GitHubEnterpriseConfig represents a configuration for a GitHub Enterprise server. + "appId": "A String", # Required. The GitHub app id of the Cloud Build app on the GitHub Enterprise server. + "createTime": "A String", # Output only. Time when the installation was associated with the project. + "displayName": "A String", # Name to display for this config. + "hostUrl": "A String", # The URL of the github enterprise host the configuration is for. + "name": "A String", # Optional. The full resource name for the GitHubEnterpriseConfig For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" + "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the GitHub Enterprise server. The VPC network must be enabled for private service connection. This should be set if the GitHub Enterprise server is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the GitHub Enterprise server will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project. + "secrets": { # GitHubEnterpriseSecrets represents the names of all necessary secrets in Secret Manager for a GitHub Enterprise server. Format is: projects//secrets/. # Names of secrets in Secret Manager. + "oauthClientIdName": "A String", # The resource name for the OAuth client ID secret in Secret Manager. + "oauthClientIdVersionName": "A String", # The resource name for the OAuth client ID secret version in Secret Manager. + "oauthSecretName": "A String", # The resource name for the OAuth secret in Secret Manager. + "oauthSecretVersionName": "A String", # The resource name for the OAuth secret secret version in Secret Manager. + "privateKeyName": "A String", # The resource name for the private key secret. + "privateKeyVersionName": "A String", # The resource name for the private key secret version. + "webhookSecretName": "A String", # The resource name for the webhook secret in Secret Manager. + "webhookSecretVersionName": "A String", # The resource name for the webhook secret secret version in Secret Manager. + }, + "sslCa": "A String", # Optional. SSL certificate to use for requests to GitHub Enterprise. + "webhookKey": "A String", # The key that should be attached to webhook calls to the ReceiveWebhook endpoint. +}+++ +list(parent, projectId=None, x__xgafv=None)+List all GitHubEnterpriseConfigs for a given project. This API is experimental. + +Args: + parent: string, Name of the parent project. For example: projects/{$project_number} or projects/{$project_id} (required) + projectId: string, ID of the project + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # RPC response object returned by ListGithubEnterpriseConfigs RPC method. + "configs": [ # A list of GitHubEnterpriseConfigs + { # GitHubEnterpriseConfig represents a configuration for a GitHub Enterprise server. + "appId": "A String", # Required. The GitHub app id of the Cloud Build app on the GitHub Enterprise server. + "createTime": "A String", # Output only. Time when the installation was associated with the project. + "displayName": "A String", # Name to display for this config. + "hostUrl": "A String", # The URL of the github enterprise host the configuration is for. + "name": "A String", # Optional. The full resource name for the GitHubEnterpriseConfig For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" + "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the GitHub Enterprise server. The VPC network must be enabled for private service connection. This should be set if the GitHub Enterprise server is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the GitHub Enterprise server will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project. + "secrets": { # GitHubEnterpriseSecrets represents the names of all necessary secrets in Secret Manager for a GitHub Enterprise server. Format is: projects//secrets/. # Names of secrets in Secret Manager. + "oauthClientIdName": "A String", # The resource name for the OAuth client ID secret in Secret Manager. + "oauthClientIdVersionName": "A String", # The resource name for the OAuth client ID secret version in Secret Manager. + "oauthSecretName": "A String", # The resource name for the OAuth secret in Secret Manager. + "oauthSecretVersionName": "A String", # The resource name for the OAuth secret secret version in Secret Manager. + "privateKeyName": "A String", # The resource name for the private key secret. + "privateKeyVersionName": "A String", # The resource name for the private key secret version. + "webhookSecretName": "A String", # The resource name for the webhook secret in Secret Manager. + "webhookSecretVersionName": "A String", # The resource name for the webhook secret secret version in Secret Manager. + }, + "sslCa": "A String", # Optional. SSL certificate to use for requests to GitHub Enterprise. + "webhookKey": "A String", # The key that should be attached to webhook calls to the ReceiveWebhook endpoint. + }, + ], +}+++ + \ No newline at end of file diff --git a/docs/dyn/cloudbuild_v1.projects.html b/docs/dyn/cloudbuild_v1.projects.html index 1ccb0cd561e..d8e4fa4d008 100644 --- a/docs/dyn/cloudbuild_v1.projects.html +++ b/docs/dyn/cloudbuild_v1.projects.html @@ -79,6 +79,11 @@patch(name, body=None, updateMask=None, x__xgafv=None)+Update an association between a GCP project and a GitHub Enterprise server. This API is experimental. + +Args: + name: string, Optional. The full resource name for the GitHubEnterpriseConfig For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" (required) + body: object, The request body. + The object takes the form of: + +{ # GitHubEnterpriseConfig represents a configuration for a GitHub Enterprise server. + "appId": "A String", # Required. The GitHub app id of the Cloud Build app on the GitHub Enterprise server. + "createTime": "A String", # Output only. Time when the installation was associated with the project. + "displayName": "A String", # Name to display for this config. + "hostUrl": "A String", # The URL of the github enterprise host the configuration is for. + "name": "A String", # Optional. The full resource name for the GitHubEnterpriseConfig For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" + "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the GitHub Enterprise server. The VPC network must be enabled for private service connection. This should be set if the GitHub Enterprise server is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the GitHub Enterprise server will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project. + "secrets": { # GitHubEnterpriseSecrets represents the names of all necessary secrets in Secret Manager for a GitHub Enterprise server. Format is: projects//secrets/. # Names of secrets in Secret Manager. + "oauthClientIdName": "A String", # The resource name for the OAuth client ID secret in Secret Manager. + "oauthClientIdVersionName": "A String", # The resource name for the OAuth client ID secret version in Secret Manager. + "oauthSecretName": "A String", # The resource name for the OAuth secret in Secret Manager. + "oauthSecretVersionName": "A String", # The resource name for the OAuth secret secret version in Secret Manager. + "privateKeyName": "A String", # The resource name for the private key secret. + "privateKeyVersionName": "A String", # The resource name for the private key secret version. + "webhookSecretName": "A String", # The resource name for the webhook secret in Secret Manager. + "webhookSecretVersionName": "A String", # The resource name for the webhook secret secret version in Secret Manager. + }, + "sslCa": "A String", # Optional. SSL certificate to use for requests to GitHub Enterprise. + "webhookKey": "A String", # The key that should be attached to webhook calls to the ReceiveWebhook endpoint. +} + + updateMask: string, Update mask for the resource. If this is set, the server will only update the fields specified in the field mask. Otherwise, a full update of the mutable resource fields will be performed. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+Instance Methods
Returns the builds Resource.
+ +Returns the githubEnterpriseConfigs Resource.
+ diff --git a/docs/dyn/cloudbuild_v1.projects.locations.builds.html b/docs/dyn/cloudbuild_v1.projects.locations.builds.html index 3e2fd20eb3d..e53a725374c 100644 --- a/docs/dyn/cloudbuild_v1.projects.locations.builds.html +++ b/docs/dyn/cloudbuild_v1.projects.locations.builds.html @@ -332,7 +332,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -574,7 +574,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -847,7 +847,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -1095,7 +1095,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. diff --git a/docs/dyn/cloudbuild_v1.projects.locations.githubEnterpriseConfigs.html b/docs/dyn/cloudbuild_v1.projects.locations.githubEnterpriseConfigs.html new file mode 100644 index 00000000000..8a0949722ad --- /dev/null +++ b/docs/dyn/cloudbuild_v1.projects.locations.githubEnterpriseConfigs.html @@ -0,0 +1,337 @@ + + + +Cloud Build API . projects . locations . githubEnterpriseConfigs
+Instance Methods
++
+close()Close httplib2 connections.
++
+create(parent, body=None, projectId=None, x__xgafv=None)Create an association between a GCP project and a GitHub Enterprise server. This API is experimental.
++
+delete(name, configId=None, projectId=None, x__xgafv=None)Delete an association between a GCP project and a GitHub Enterprise server. This API is experimental.
++
+get(name, configId=None, projectId=None, x__xgafv=None)Retrieve a GitHubEnterpriseConfig. This API is experimental.
++
+list(parent, projectId=None, x__xgafv=None)List all GitHubEnterpriseConfigs for a given project. This API is experimental.
++
+patch(name, body=None, updateMask=None, x__xgafv=None)Update an association between a GCP project and a GitHub Enterprise server. This API is experimental.
+Method Details
+++ +close()+Close httplib2 connections.+++ +create(parent, body=None, projectId=None, x__xgafv=None)+Create an association between a GCP project and a GitHub Enterprise server. This API is experimental. + +Args: + parent: string, Name of the parent project. For example: projects/{$project_number} or projects/{$project_id} (required) + body: object, The request body. + The object takes the form of: + +{ # GitHubEnterpriseConfig represents a configuration for a GitHub Enterprise server. + "appId": "A String", # Required. The GitHub app id of the Cloud Build app on the GitHub Enterprise server. + "createTime": "A String", # Output only. Time when the installation was associated with the project. + "displayName": "A String", # Name to display for this config. + "hostUrl": "A String", # The URL of the github enterprise host the configuration is for. + "name": "A String", # Optional. The full resource name for the GitHubEnterpriseConfig For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" + "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the GitHub Enterprise server. The VPC network must be enabled for private service connection. This should be set if the GitHub Enterprise server is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the GitHub Enterprise server will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project. + "secrets": { # GitHubEnterpriseSecrets represents the names of all necessary secrets in Secret Manager for a GitHub Enterprise server. Format is: projects//secrets/. # Names of secrets in Secret Manager. + "oauthClientIdName": "A String", # The resource name for the OAuth client ID secret in Secret Manager. + "oauthClientIdVersionName": "A String", # The resource name for the OAuth client ID secret version in Secret Manager. + "oauthSecretName": "A String", # The resource name for the OAuth secret in Secret Manager. + "oauthSecretVersionName": "A String", # The resource name for the OAuth secret secret version in Secret Manager. + "privateKeyName": "A String", # The resource name for the private key secret. + "privateKeyVersionName": "A String", # The resource name for the private key secret version. + "webhookSecretName": "A String", # The resource name for the webhook secret in Secret Manager. + "webhookSecretVersionName": "A String", # The resource name for the webhook secret secret version in Secret Manager. + }, + "sslCa": "A String", # Optional. SSL certificate to use for requests to GitHub Enterprise. + "webhookKey": "A String", # The key that should be attached to webhook calls to the ReceiveWebhook endpoint. +} + + projectId: string, ID of the project. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +delete(name, configId=None, projectId=None, x__xgafv=None)+Delete an association between a GCP project and a GitHub Enterprise server. This API is experimental. + +Args: + name: string, This field should contain the name of the enterprise config resource. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" (required) + configId: string, Unique identifier of the `GitHubEnterpriseConfig` + projectId: string, ID of the project + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +get(name, configId=None, projectId=None, x__xgafv=None)+Retrieve a GitHubEnterpriseConfig. This API is experimental. + +Args: + name: string, This field should contain the name of the enterprise config resource. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" (required) + configId: string, Unique identifier of the `GitHubEnterpriseConfig` + projectId: string, ID of the project + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # GitHubEnterpriseConfig represents a configuration for a GitHub Enterprise server. + "appId": "A String", # Required. The GitHub app id of the Cloud Build app on the GitHub Enterprise server. + "createTime": "A String", # Output only. Time when the installation was associated with the project. + "displayName": "A String", # Name to display for this config. + "hostUrl": "A String", # The URL of the github enterprise host the configuration is for. + "name": "A String", # Optional. The full resource name for the GitHubEnterpriseConfig For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" + "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the GitHub Enterprise server. The VPC network must be enabled for private service connection. This should be set if the GitHub Enterprise server is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the GitHub Enterprise server will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project. + "secrets": { # GitHubEnterpriseSecrets represents the names of all necessary secrets in Secret Manager for a GitHub Enterprise server. Format is: projects//secrets/. # Names of secrets in Secret Manager. + "oauthClientIdName": "A String", # The resource name for the OAuth client ID secret in Secret Manager. + "oauthClientIdVersionName": "A String", # The resource name for the OAuth client ID secret version in Secret Manager. + "oauthSecretName": "A String", # The resource name for the OAuth secret in Secret Manager. + "oauthSecretVersionName": "A String", # The resource name for the OAuth secret secret version in Secret Manager. + "privateKeyName": "A String", # The resource name for the private key secret. + "privateKeyVersionName": "A String", # The resource name for the private key secret version. + "webhookSecretName": "A String", # The resource name for the webhook secret in Secret Manager. + "webhookSecretVersionName": "A String", # The resource name for the webhook secret secret version in Secret Manager. + }, + "sslCa": "A String", # Optional. SSL certificate to use for requests to GitHub Enterprise. + "webhookKey": "A String", # The key that should be attached to webhook calls to the ReceiveWebhook endpoint. +}+++ +list(parent, projectId=None, x__xgafv=None)+List all GitHubEnterpriseConfigs for a given project. This API is experimental. + +Args: + parent: string, Name of the parent project. For example: projects/{$project_number} or projects/{$project_id} (required) + projectId: string, ID of the project + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # RPC response object returned by ListGithubEnterpriseConfigs RPC method. + "configs": [ # A list of GitHubEnterpriseConfigs + { # GitHubEnterpriseConfig represents a configuration for a GitHub Enterprise server. + "appId": "A String", # Required. The GitHub app id of the Cloud Build app on the GitHub Enterprise server. + "createTime": "A String", # Output only. Time when the installation was associated with the project. + "displayName": "A String", # Name to display for this config. + "hostUrl": "A String", # The URL of the github enterprise host the configuration is for. + "name": "A String", # Optional. The full resource name for the GitHubEnterpriseConfig For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" + "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the GitHub Enterprise server. The VPC network must be enabled for private service connection. This should be set if the GitHub Enterprise server is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the GitHub Enterprise server will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project. + "secrets": { # GitHubEnterpriseSecrets represents the names of all necessary secrets in Secret Manager for a GitHub Enterprise server. Format is: projects//secrets/. # Names of secrets in Secret Manager. + "oauthClientIdName": "A String", # The resource name for the OAuth client ID secret in Secret Manager. + "oauthClientIdVersionName": "A String", # The resource name for the OAuth client ID secret version in Secret Manager. + "oauthSecretName": "A String", # The resource name for the OAuth secret in Secret Manager. + "oauthSecretVersionName": "A String", # The resource name for the OAuth secret secret version in Secret Manager. + "privateKeyName": "A String", # The resource name for the private key secret. + "privateKeyVersionName": "A String", # The resource name for the private key secret version. + "webhookSecretName": "A String", # The resource name for the webhook secret in Secret Manager. + "webhookSecretVersionName": "A String", # The resource name for the webhook secret secret version in Secret Manager. + }, + "sslCa": "A String", # Optional. SSL certificate to use for requests to GitHub Enterprise. + "webhookKey": "A String", # The key that should be attached to webhook calls to the ReceiveWebhook endpoint. + }, + ], +}+++ + \ No newline at end of file diff --git a/docs/dyn/cloudbuild_v1.projects.locations.html b/docs/dyn/cloudbuild_v1.projects.locations.html index 19eeb5511f3..a977d5c0407 100644 --- a/docs/dyn/cloudbuild_v1.projects.locations.html +++ b/docs/dyn/cloudbuild_v1.projects.locations.html @@ -79,6 +79,11 @@patch(name, body=None, updateMask=None, x__xgafv=None)+Update an association between a GCP project and a GitHub Enterprise server. This API is experimental. + +Args: + name: string, Optional. The full resource name for the GitHubEnterpriseConfig For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" (required) + body: object, The request body. + The object takes the form of: + +{ # GitHubEnterpriseConfig represents a configuration for a GitHub Enterprise server. + "appId": "A String", # Required. The GitHub app id of the Cloud Build app on the GitHub Enterprise server. + "createTime": "A String", # Output only. Time when the installation was associated with the project. + "displayName": "A String", # Name to display for this config. + "hostUrl": "A String", # The URL of the github enterprise host the configuration is for. + "name": "A String", # Optional. The full resource name for the GitHubEnterpriseConfig For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" + "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the GitHub Enterprise server. The VPC network must be enabled for private service connection. This should be set if the GitHub Enterprise server is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the GitHub Enterprise server will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project. + "secrets": { # GitHubEnterpriseSecrets represents the names of all necessary secrets in Secret Manager for a GitHub Enterprise server. Format is: projects//secrets/. # Names of secrets in Secret Manager. + "oauthClientIdName": "A String", # The resource name for the OAuth client ID secret in Secret Manager. + "oauthClientIdVersionName": "A String", # The resource name for the OAuth client ID secret version in Secret Manager. + "oauthSecretName": "A String", # The resource name for the OAuth secret in Secret Manager. + "oauthSecretVersionName": "A String", # The resource name for the OAuth secret secret version in Secret Manager. + "privateKeyName": "A String", # The resource name for the private key secret. + "privateKeyVersionName": "A String", # The resource name for the private key secret version. + "webhookSecretName": "A String", # The resource name for the webhook secret in Secret Manager. + "webhookSecretVersionName": "A String", # The resource name for the webhook secret secret version in Secret Manager. + }, + "sslCa": "A String", # Optional. SSL certificate to use for requests to GitHub Enterprise. + "webhookKey": "A String", # The key that should be attached to webhook calls to the ReceiveWebhook endpoint. +} + + updateMask: string, Update mask for the resource. If this is set, the server will only update the fields specified in the field mask. Otherwise, a full update of the mutable resource fields will be performed. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+Instance Methods
Returns the builds Resource.
+ +Returns the githubEnterpriseConfigs Resource.
+ diff --git a/docs/dyn/cloudbuild_v1.projects.locations.triggers.html b/docs/dyn/cloudbuild_v1.projects.locations.triggers.html index 5c2898e074f..dfc1f130c60 100644 --- a/docs/dyn/cloudbuild_v1.projects.locations.triggers.html +++ b/docs/dyn/cloudbuild_v1.projects.locations.triggers.html @@ -331,7 +331,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -349,7 +349,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -379,6 +380,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. @@ -632,7 +634,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -650,7 +652,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -680,6 +683,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. @@ -961,7 +965,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -979,7 +983,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -1009,6 +1014,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. @@ -1274,7 +1280,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -1292,7 +1298,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -1322,6 +1329,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. @@ -1592,7 +1600,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -1610,7 +1618,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -1640,6 +1649,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. @@ -1894,7 +1904,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -1912,7 +1922,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -1942,6 +1953,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. diff --git a/docs/dyn/cloudbuild_v1.projects.triggers.html b/docs/dyn/cloudbuild_v1.projects.triggers.html index 14c7e600afd..5c499c3acb2 100644 --- a/docs/dyn/cloudbuild_v1.projects.triggers.html +++ b/docs/dyn/cloudbuild_v1.projects.triggers.html @@ -331,7 +331,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -349,7 +349,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -379,6 +380,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. @@ -632,7 +634,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -650,7 +652,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -680,6 +683,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. @@ -961,7 +965,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -979,7 +983,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -1009,6 +1014,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. @@ -1274,7 +1280,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -1292,7 +1298,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -1322,6 +1329,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. @@ -1593,7 +1601,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -1611,7 +1619,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -1641,6 +1650,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. @@ -1893,7 +1903,7 @@Method Details
"A String", ], "timeout": "A String", # Amount of time that this build should be allowed to run, to second granularity. If this amount of time elapses, work on the build will cease and the build status will be `TIMEOUT`. `timeout` starts ticking from `startTime`. Default time is ten minutes. - "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. If the build does not specify source or images, these keys will not be included. + "timing": { # Output only. Stores timing information for phases of the build. Valid keys are: * BUILD: time to execute all build steps. * PUSH: time to push all specified images. * FETCHSOURCE: time to fetch source. * SETUPBUILD: time to set up build. If the build does not specify source or images, these keys will not be included. "a_key": { # Start and end times for a build execution phase. "endTime": "A String", # End of time span. "startTime": "A String", # Start of time span. @@ -1911,7 +1921,8 @@Method Details
"disabled": True or False, # If true, the trigger will never automatically execute a build. "filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml). "filter": "A String", # Optional. A Common Expression Language string. - "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. This message is experimental. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "github": { # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. # GitHubEventsConfig describes the configuration of a trigger that creates a build whenever a GitHub event is received. Mutually exclusive with `trigger_template`. + "enterpriseConfigResourceName": "A String", # Optional. The resource name of the github enterprise config that should be applied to this installation. For example: "projects/{$project_id}/githubEnterpriseConfig/{$config_id}" "installationId": "A String", # The installationID that emits the GitHub event. "name": "A String", # Name of the repository. For example: The name for https://github.com/googlecloudplatform/cloud-builders is "cloud-builders". "owner": "A String", # Owner of the repository. For example: The owner for https://github.com/googlecloudplatform/cloud-builders is "googlecloudplatform". @@ -1941,6 +1952,7 @@Method Details
"topic": "A String", # The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. }, "resourceName": "A String", # The `Trigger` name with format: `projects/{project}/locations/{location}/triggers/{trigger}`, where {trigger} is a unique identifier generated by the service. + "serviceAccount": "A String", # Optional. The service account used for all user-controlled operations including UpdateBuildTrigger, RunBuildTrigger, CreateBuild, and CancelBuild. If no service account is set, then the standard Cloud Build service account ([PROJECT_NUM]@system.gserviceaccount.com) will be used instead. Format: `projects/{PROJECT_ID}/serviceAccounts/{ACCOUNT_ID_OR_EMAIL}` "sourceToBuild": { # GitRepoSource describes a repo and ref of a code repository. # The repo and ref of the repository from which to build. This field is used only for those triggers that do not respond to SCM events. Triggers that respond to such events build source at whatever commit caused the event. This field is currently only used by Webhook, Pub/Sub, Manual, and Cron triggers. "ref": "A String", # The branch or tag to use. Must start with "refs/" (required). "repoType": "A String", # See RepoType below. diff --git a/docs/dyn/cloudbuild_v1.v1.html b/docs/dyn/cloudbuild_v1.v1.html new file mode 100644 index 00000000000..ed124eb0970 --- /dev/null +++ b/docs/dyn/cloudbuild_v1.v1.html @@ -0,0 +1,120 @@ + + + +Cloud Build API . v1
+Instance Methods
++
+close()Close httplib2 connections.
++
+webhook(body=None, webhookKey=None, x__xgafv=None)ReceiveWebhook is called when the API receives a GitHub webhook.
+Method Details
+++ +close()+Close httplib2 connections.+++ + \ No newline at end of file diff --git a/docs/dyn/cloudchannel_v1.accounts.channelPartnerLinks.customers.html b/docs/dyn/cloudchannel_v1.accounts.channelPartnerLinks.customers.html index b0f1ab64a5e..52074354f98 100644 --- a/docs/dyn/cloudchannel_v1.accounts.channelPartnerLinks.customers.html +++ b/docs/dyn/cloudchannel_v1.accounts.channelPartnerLinks.customers.html @@ -86,6 +86,9 @@webhook(body=None, webhookKey=None, x__xgafv=None)+ReceiveWebhook is called when the API receives a GitHub webhook. + +Args: + body: object, The request body. + The object takes the form of: + +{ # Message that represents an arbitrary HTTP body. It should only be used for payload formats that can't be represented as JSON, such as raw binary or an HTML page. This message can be used both in streaming and non-streaming API methods in the request as well as the response. It can be used as a top-level request field, which is convenient if one wants to extract parameters from either the URL or HTTP template into the request fields and also want access to the raw HTTP body. Example: message GetResourceRequest { // A unique request id. string request_id = 1; // The raw HTTP body is bound to this field. google.api.HttpBody http_body = 2; } service ResourceService { rpc GetResource(GetResourceRequest) returns (google.api.HttpBody); rpc UpdateResource(google.api.HttpBody) returns (google.protobuf.Empty); } Example with streaming methods: service CaldavService { rpc GetCalendar(stream google.api.HttpBody) returns (stream google.api.HttpBody); rpc UpdateCalendar(stream google.api.HttpBody) returns (stream google.api.HttpBody); } Use of this type only changes how the request and response bodies are handled, all other features will continue to work unchanged. + "contentType": "A String", # The HTTP Content-Type header value specifying the content type of the body. + "data": "A String", # The HTTP request/response body as raw binary. + "extensions": [ # Application specific response metadata. Must be set in the first response for streaming APIs. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], +} + + webhookKey: string, For GitHub Enterprise webhooks, this key is used to associate the webhook request with the GitHubEnterpriseConfig to use for validation. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`. +}+Instance Methods
Returns the requested Customer resource. Possible error codes: * PERMISSION_DENIED: The reseller account making the request is different from the reseller account in the API request. * INVALID_ARGUMENT: Required request parameters are missing or invalid. * NOT_FOUND: The customer resource doesn't exist. Usually the result of an invalid name parameter. Return value: The Customer resource.
++
+import_(parent, body=None, x__xgafv=None)Imports a Customer from the Cloud Identity associated with the provided Cloud Identity ID or domain before a TransferEntitlements call. If a linked Customer already exists and overwrite_if_exists is true, it will update that Customer's data. Possible error codes: * PERMISSION_DENIED: The reseller account making the request is different from the reseller account in the API request. * NOT_FOUND: Cloud Identity doesn't exist or was deleted. * INVALID_ARGUMENT: Required parameters are missing, or the auth_token is expired or invalid. * ALREADY_EXISTS: A customer already exists and has conflicting critical fields. Requires an overwrite. Return value: The Customer.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)List Customers. Possible error codes: * PERMISSION_DENIED: The reseller account making the request is different from the reseller account in the API request. * INVALID_ARGUMENT: Required request parameters are missing or invalid. Return value: List of Customers, or an empty list if there are no customers.
@@ -305,6 +308,84 @@Method Details
}
import_(parent, body=None, x__xgafv=None)
+ Imports a Customer from the Cloud Identity associated with the provided Cloud Identity ID or domain before a TransferEntitlements call. If a linked Customer already exists and overwrite_if_exists is true, it will update that Customer's data. Possible error codes: * PERMISSION_DENIED: The reseller account making the request is different from the reseller account in the API request. * NOT_FOUND: Cloud Identity doesn't exist or was deleted. * INVALID_ARGUMENT: Required parameters are missing, or the auth_token is expired or invalid. * ALREADY_EXISTS: A customer already exists and has conflicting critical fields. Requires an overwrite. Return value: The Customer.
+
+Args:
+ parent: string, Required. The resource name of the reseller's account. Parent takes the format: accounts/{account_id} or accounts/{account_id}/channelPartnerLinks/{channel_partner_id} (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for CloudChannelService.ImportCustomer
+ "authToken": "A String", # Optional. The super admin of the resold customer generates this token to authorize a reseller to access their Cloud Identity and purchase entitlements on their behalf. You can omit this token after authorization. See https://support.google.com/a/answer/7643790 for more details.
+ "channelPartnerId": "A String", # Optional. Cloud Identity ID of a channel partner who will be the direct reseller for the customer's order. This field is required for 2-tier transfer scenarios and can be provided via the request Parent binding as well.
+ "cloudIdentityId": "A String", # Required. Customer's Cloud Identity ID
+ "customer": "A String", # Optional. Specifies the customer that will receive imported Cloud Identity information. Format: accounts/{account_id}/customers/{customer_id}
+ "domain": "A String", # Required. Customer domain.
+ "overwriteIfExists": True or False, # Required. Choose to overwrite an existing customer if found. This must be set to true if there is an existing customer with a conflicting region code or domain.
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Entity representing a customer of a reseller or distributor.
+ "alternateEmail": "A String", # Secondary contact email. You need to provide an alternate email to create different domains if a primary contact email already exists. Users will receive a notification with credentials when you create an admin.google.com account. Secondary emails are also recovery email addresses. Alternate emails are optional when you create Team customers.
+ "channelPartnerId": "A String", # Cloud Identity ID of the customer's channel partner. Populated only if a channel partner exists for this customer.
+ "cloudIdentityId": "A String", # Output only. The customer's Cloud Identity ID if the customer has a Cloud Identity resource.
+ "cloudIdentityInfo": { # Cloud Identity information for the Cloud Channel Customer. # Output only. Cloud Identity information for the customer. Populated only if a Cloud Identity account exists for this customer.
+ "adminConsoleUri": "A String", # Output only. URI of Customer's Admin console dashboard.
+ "alternateEmail": "A String", # The alternate email.
+ "customerType": "A String", # CustomerType indicates verification type needed for using services.
+ "eduData": { # Required Edu Attributes # Edu information about the customer.
+ "instituteSize": "A String", # Size of the institute.
+ "instituteType": "A String", # Designated institute type of customer.
+ "website": "A String", # Web address for the edu customer's institution.
+ },
+ "isDomainVerified": True or False, # Output only. Whether the domain is verified. This field is not returned for a Customer's cloud_identity_info resource. Partners can use the domains.get() method of the Workspace SDK's Directory API, or listen to the PRIMARY_DOMAIN_VERIFIED Pub/Sub event in to track domain verification of their resolve Workspace customers.
+ "languageCode": "A String", # Language code.
+ "phoneNumber": "A String", # Phone number associated with the Cloud Identity.
+ "primaryDomain": "A String", # Output only. The primary domain name.
+ },
+ "createTime": "A String", # Output only. Time when the customer was created.
+ "domain": "A String", # Required. The customer's primary domain. Must match the primary contact email's domain.
+ "languageCode": "A String", # Optional. The BCP-47 language code, such as "en-US" or "sr-Latn". For more information, see https://www.unicode.org/reports/tr35/#Unicode_locale_identifier.
+ "name": "A String", # Output only. Resource name of the customer. Format: accounts/{account_id}/customers/{customer_id}
+ "orgDisplayName": "A String", # Required. Name of the organization that the customer entity represents.
+ "orgPostalAddress": { # Represents a postal address, e.g. for postal delivery or payments addresses. Given a postal address, a postal service can deliver items to a premise, P.O. Box or similar. It is not intended to model geographical locations (roads, towns, mountains). In typical usage an address would be created via user input or from importing existing data, depending on the type of process. Advice on address input / editing: - Use an i18n-ready address widget such as https://github.com/google/libaddressinput) - Users should not be presented with UI elements for input or editing of fields outside countries where that field is used. For more guidance on how to use this schema, please see: https://support.google.com/business/answer/6397478 # Required. The organization address for the customer. To enforce US laws and embargoes, we require a region and zip code. You must provide valid addresses for every customer. To set the customer's language, use the Customer-level language code.
+ "addressLines": [ # Unstructured address lines describing the lower levels of an address. Because values in address_lines do not have type information and may sometimes contain multiple values in a single field (e.g. "Austin, TX"), it is important that the line order is clear. The order of address lines should be "envelope order" for the country/region of the address. In places where this can vary (e.g. Japan), address_language is used to make it explicit (e.g. "ja" for large-to-small ordering and "ja-Latn" or "en" for small-to-large). This way, the most specific line of an address can be selected based on the language. The minimum permitted structural representation of an address consists of a region_code with all remaining information placed in the address_lines. It would be possible to format such an address very approximately without geocoding, but no semantic reasoning could be made about any of the address components until it was at least partially resolved. Creating an address only containing a region_code and address_lines, and then geocoding is the recommended way to handle completely unstructured addresses (as opposed to guessing which parts of the address should be localities or administrative areas).
+ "A String",
+ ],
+ "administrativeArea": "A String", # Optional. Highest administrative subdivision which is used for postal addresses of a country or region. For example, this can be a state, a province, an oblast, or a prefecture. Specifically, for Spain this is the province and not the autonomous community (e.g. "Barcelona" and not "Catalonia"). Many countries don't use an administrative area in postal addresses. E.g. in Switzerland this should be left unpopulated.
+ "languageCode": "A String", # Optional. BCP-47 language code of the contents of this address (if known). This is often the UI language of the input form or is expected to match one of the languages used in the address' country/region, or their transliterated equivalents. This can affect formatting in certain countries, but is not critical to the correctness of the data and will never affect any validation or other non-formatting related operations. If this value is not known, it should be omitted (rather than specifying a possibly incorrect default). Examples: "zh-Hant", "ja", "ja-Latn", "en".
+ "locality": "A String", # Optional. Generally refers to the city/town portion of the address. Examples: US city, IT comune, UK post town. In regions of the world where localities are not well defined or do not fit into this structure well, leave locality empty and use address_lines.
+ "organization": "A String", # Optional. The name of the organization at the address.
+ "postalCode": "A String", # Optional. Postal code of the address. Not all countries use or require postal codes to be present, but where they are used, they may trigger additional validation with other parts of the address (e.g. state/zip validation in the U.S.A.).
+ "recipients": [ # Optional. The recipient at the address. This field may, under certain circumstances, contain multiline information. For example, it might contain "care of" information.
+ "A String",
+ ],
+ "regionCode": "A String", # Required. CLDR region code of the country/region of the address. This is never inferred and it is up to the user to ensure the value is correct. See http://cldr.unicode.org/ and http://www.unicode.org/cldr/charts/30/supplemental/territory_information.html for details. Example: "CH" for Switzerland.
+ "revision": 42, # The schema revision of the `PostalAddress`. This must be set to 0, which is the latest revision. All new revisions **must** be backward compatible with old revisions.
+ "sortingCode": "A String", # Optional. Additional, country-specific, sorting code. This is not used in most regions. Where it is used, the value is either a string like "CEDEX", optionally followed by a number (e.g. "CEDEX 7"), or just a number alone, representing the "sector code" (Jamaica), "delivery area indicator" (Malawi) or "post office indicator" (e.g. Côte d'Ivoire).
+ "sublocality": "A String", # Optional. Sublocality of the address. For example, this can be neighborhoods, boroughs, districts.
+ },
+ "primaryContactInfo": { # Contact information for a customer account. # Primary contact info.
+ "displayName": "A String", # Output only. The customer account contact's display name, formatted as a combination of the customer's first and last name.
+ "email": "A String", # The customer account's contact email. Required for entitlements that create admin.google.com accounts, and serves as the customer's username for those accounts. Use this email to invite Team customers.
+ "firstName": "A String", # The customer account contact's first name. Optional for Team customers.
+ "lastName": "A String", # The customer account contact's last name. Optional for Team customers.
+ "phone": "A String", # The customer account's contact phone number.
+ "title": "A String", # Optional. The customer account contact's job title.
+ },
+ "updateTime": "A String", # Output only. Time when the customer was updated.
+}
+list(parent, pageSize=None, pageToken=None, x__xgafv=None)
List Customers. Possible error codes: * PERMISSION_DENIED: The reseller account making the request is different from the reseller account in the API request. * INVALID_ARGUMENT: Required request parameters are missing or invalid. Return value: List of Customers, or an empty list if there are no customers. diff --git a/docs/dyn/cloudchannel_v1.accounts.customers.html b/docs/dyn/cloudchannel_v1.accounts.customers.html index 77437554299..211df65c47b 100644 --- a/docs/dyn/cloudchannel_v1.accounts.customers.html +++ b/docs/dyn/cloudchannel_v1.accounts.customers.html @@ -91,6 +91,9 @@Instance Methods
Returns the requested Customer resource. Possible error codes: * PERMISSION_DENIED: The reseller account making the request is different from the reseller account in the API request. * INVALID_ARGUMENT: Required request parameters are missing or invalid. * NOT_FOUND: The customer resource doesn't exist. Usually the result of an invalid name parameter. Return value: The Customer resource.
++
+import_(parent, body=None, x__xgafv=None)Imports a Customer from the Cloud Identity associated with the provided Cloud Identity ID or domain before a TransferEntitlements call. If a linked Customer already exists and overwrite_if_exists is true, it will update that Customer's data. Possible error codes: * PERMISSION_DENIED: The reseller account making the request is different from the reseller account in the API request. * NOT_FOUND: Cloud Identity doesn't exist or was deleted. * INVALID_ARGUMENT: Required parameters are missing, or the auth_token is expired or invalid. * ALREADY_EXISTS: A customer already exists and has conflicting critical fields. Requires an overwrite. Return value: The Customer.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)List Customers. Possible error codes: * PERMISSION_DENIED: The reseller account making the request is different from the reseller account in the API request. * INVALID_ARGUMENT: Required request parameters are missing or invalid. Return value: List of Customers, or an empty list if there are no customers.
@@ -331,6 +334,84 @@Method Details
}
import_(parent, body=None, x__xgafv=None)
+ Imports a Customer from the Cloud Identity associated with the provided Cloud Identity ID or domain before a TransferEntitlements call. If a linked Customer already exists and overwrite_if_exists is true, it will update that Customer's data. Possible error codes: * PERMISSION_DENIED: The reseller account making the request is different from the reseller account in the API request. * NOT_FOUND: Cloud Identity doesn't exist or was deleted. * INVALID_ARGUMENT: Required parameters are missing, or the auth_token is expired or invalid. * ALREADY_EXISTS: A customer already exists and has conflicting critical fields. Requires an overwrite. Return value: The Customer.
+
+Args:
+ parent: string, Required. The resource name of the reseller's account. Parent takes the format: accounts/{account_id} or accounts/{account_id}/channelPartnerLinks/{channel_partner_id} (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for CloudChannelService.ImportCustomer
+ "authToken": "A String", # Optional. The super admin of the resold customer generates this token to authorize a reseller to access their Cloud Identity and purchase entitlements on their behalf. You can omit this token after authorization. See https://support.google.com/a/answer/7643790 for more details.
+ "channelPartnerId": "A String", # Optional. Cloud Identity ID of a channel partner who will be the direct reseller for the customer's order. This field is required for 2-tier transfer scenarios and can be provided via the request Parent binding as well.
+ "cloudIdentityId": "A String", # Required. Customer's Cloud Identity ID
+ "customer": "A String", # Optional. Specifies the customer that will receive imported Cloud Identity information. Format: accounts/{account_id}/customers/{customer_id}
+ "domain": "A String", # Required. Customer domain.
+ "overwriteIfExists": True or False, # Required. Choose to overwrite an existing customer if found. This must be set to true if there is an existing customer with a conflicting region code or domain.
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Entity representing a customer of a reseller or distributor.
+ "alternateEmail": "A String", # Secondary contact email. You need to provide an alternate email to create different domains if a primary contact email already exists. Users will receive a notification with credentials when you create an admin.google.com account. Secondary emails are also recovery email addresses. Alternate emails are optional when you create Team customers.
+ "channelPartnerId": "A String", # Cloud Identity ID of the customer's channel partner. Populated only if a channel partner exists for this customer.
+ "cloudIdentityId": "A String", # Output only. The customer's Cloud Identity ID if the customer has a Cloud Identity resource.
+ "cloudIdentityInfo": { # Cloud Identity information for the Cloud Channel Customer. # Output only. Cloud Identity information for the customer. Populated only if a Cloud Identity account exists for this customer.
+ "adminConsoleUri": "A String", # Output only. URI of Customer's Admin console dashboard.
+ "alternateEmail": "A String", # The alternate email.
+ "customerType": "A String", # CustomerType indicates verification type needed for using services.
+ "eduData": { # Required Edu Attributes # Edu information about the customer.
+ "instituteSize": "A String", # Size of the institute.
+ "instituteType": "A String", # Designated institute type of customer.
+ "website": "A String", # Web address for the edu customer's institution.
+ },
+ "isDomainVerified": True or False, # Output only. Whether the domain is verified. This field is not returned for a Customer's cloud_identity_info resource. Partners can use the domains.get() method of the Workspace SDK's Directory API, or listen to the PRIMARY_DOMAIN_VERIFIED Pub/Sub event in to track domain verification of their resolve Workspace customers.
+ "languageCode": "A String", # Language code.
+ "phoneNumber": "A String", # Phone number associated with the Cloud Identity.
+ "primaryDomain": "A String", # Output only. The primary domain name.
+ },
+ "createTime": "A String", # Output only. Time when the customer was created.
+ "domain": "A String", # Required. The customer's primary domain. Must match the primary contact email's domain.
+ "languageCode": "A String", # Optional. The BCP-47 language code, such as "en-US" or "sr-Latn". For more information, see https://www.unicode.org/reports/tr35/#Unicode_locale_identifier.
+ "name": "A String", # Output only. Resource name of the customer. Format: accounts/{account_id}/customers/{customer_id}
+ "orgDisplayName": "A String", # Required. Name of the organization that the customer entity represents.
+ "orgPostalAddress": { # Represents a postal address, e.g. for postal delivery or payments addresses. Given a postal address, a postal service can deliver items to a premise, P.O. Box or similar. It is not intended to model geographical locations (roads, towns, mountains). In typical usage an address would be created via user input or from importing existing data, depending on the type of process. Advice on address input / editing: - Use an i18n-ready address widget such as https://github.com/google/libaddressinput) - Users should not be presented with UI elements for input or editing of fields outside countries where that field is used. For more guidance on how to use this schema, please see: https://support.google.com/business/answer/6397478 # Required. The organization address for the customer. To enforce US laws and embargoes, we require a region and zip code. You must provide valid addresses for every customer. To set the customer's language, use the Customer-level language code.
+ "addressLines": [ # Unstructured address lines describing the lower levels of an address. Because values in address_lines do not have type information and may sometimes contain multiple values in a single field (e.g. "Austin, TX"), it is important that the line order is clear. The order of address lines should be "envelope order" for the country/region of the address. In places where this can vary (e.g. Japan), address_language is used to make it explicit (e.g. "ja" for large-to-small ordering and "ja-Latn" or "en" for small-to-large). This way, the most specific line of an address can be selected based on the language. The minimum permitted structural representation of an address consists of a region_code with all remaining information placed in the address_lines. It would be possible to format such an address very approximately without geocoding, but no semantic reasoning could be made about any of the address components until it was at least partially resolved. Creating an address only containing a region_code and address_lines, and then geocoding is the recommended way to handle completely unstructured addresses (as opposed to guessing which parts of the address should be localities or administrative areas).
+ "A String",
+ ],
+ "administrativeArea": "A String", # Optional. Highest administrative subdivision which is used for postal addresses of a country or region. For example, this can be a state, a province, an oblast, or a prefecture. Specifically, for Spain this is the province and not the autonomous community (e.g. "Barcelona" and not "Catalonia"). Many countries don't use an administrative area in postal addresses. E.g. in Switzerland this should be left unpopulated.
+ "languageCode": "A String", # Optional. BCP-47 language code of the contents of this address (if known). This is often the UI language of the input form or is expected to match one of the languages used in the address' country/region, or their transliterated equivalents. This can affect formatting in certain countries, but is not critical to the correctness of the data and will never affect any validation or other non-formatting related operations. If this value is not known, it should be omitted (rather than specifying a possibly incorrect default). Examples: "zh-Hant", "ja", "ja-Latn", "en".
+ "locality": "A String", # Optional. Generally refers to the city/town portion of the address. Examples: US city, IT comune, UK post town. In regions of the world where localities are not well defined or do not fit into this structure well, leave locality empty and use address_lines.
+ "organization": "A String", # Optional. The name of the organization at the address.
+ "postalCode": "A String", # Optional. Postal code of the address. Not all countries use or require postal codes to be present, but where they are used, they may trigger additional validation with other parts of the address (e.g. state/zip validation in the U.S.A.).
+ "recipients": [ # Optional. The recipient at the address. This field may, under certain circumstances, contain multiline information. For example, it might contain "care of" information.
+ "A String",
+ ],
+ "regionCode": "A String", # Required. CLDR region code of the country/region of the address. This is never inferred and it is up to the user to ensure the value is correct. See http://cldr.unicode.org/ and http://www.unicode.org/cldr/charts/30/supplemental/territory_information.html for details. Example: "CH" for Switzerland.
+ "revision": 42, # The schema revision of the `PostalAddress`. This must be set to 0, which is the latest revision. All new revisions **must** be backward compatible with old revisions.
+ "sortingCode": "A String", # Optional. Additional, country-specific, sorting code. This is not used in most regions. Where it is used, the value is either a string like "CEDEX", optionally followed by a number (e.g. "CEDEX 7"), or just a number alone, representing the "sector code" (Jamaica), "delivery area indicator" (Malawi) or "post office indicator" (e.g. Côte d'Ivoire).
+ "sublocality": "A String", # Optional. Sublocality of the address. For example, this can be neighborhoods, boroughs, districts.
+ },
+ "primaryContactInfo": { # Contact information for a customer account. # Primary contact info.
+ "displayName": "A String", # Output only. The customer account contact's display name, formatted as a combination of the customer's first and last name.
+ "email": "A String", # The customer account's contact email. Required for entitlements that create admin.google.com accounts, and serves as the customer's username for those accounts. Use this email to invite Team customers.
+ "firstName": "A String", # The customer account contact's first name. Optional for Team customers.
+ "lastName": "A String", # The customer account contact's last name. Optional for Team customers.
+ "phone": "A String", # The customer account's contact phone number.
+ "title": "A String", # Optional. The customer account contact's job title.
+ },
+ "updateTime": "A String", # Output only. Time when the customer was updated.
+}
+list(parent, pageSize=None, pageToken=None, x__xgafv=None)
List Customers. Possible error codes: * PERMISSION_DENIED: The reseller account making the request is different from the reseller account in the API request. * INVALID_ARGUMENT: Required request parameters are missing or invalid. Return value: List of Customers, or an empty list if there are no customers. diff --git a/docs/dyn/cloudkms_v1.projects.locations.html b/docs/dyn/cloudkms_v1.projects.locations.html index 8e3d8f957be..0c07cbb065b 100644 --- a/docs/dyn/cloudkms_v1.projects.locations.html +++ b/docs/dyn/cloudkms_v1.projects.locations.html @@ -82,6 +82,9 @@Instance Methods
Close httplib2 connections.
++
+generateRandomBytes(location, body=None, x__xgafv=None)Generate random bytes using the Cloud KMS randomness source in the provided location.
Gets information about a location.
@@ -97,6 +100,34 @@Method Details
Close httplib2 connections.
generateRandomBytes(location, body=None, x__xgafv=None)
+ Generate random bytes using the Cloud KMS randomness source in the provided location.
+
+Args:
+ location: string, The project-specific location in which to generate random bytes. For example, "projects/my-project/locations/us-central1". (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for KeyManagementService.GenerateRandomBytes.
+ "lengthBytes": 42, # The length in bytes of the amount of randomness to retrieve. Minimum 8 bytes, maximum 1024 bytes.
+ "protectionLevel": "A String", # The ProtectionLevel to use when generating the random data. Defaults to SOFTWARE.
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response message for KeyManagementService.GenerateRandomBytes.
+ "data": "A String", # The generated data.
+ "dataCrc32c": "A String", # Integrity verification field. A CRC32C checksum of the returned GenerateRandomBytesResponse.data. An integrity check of GenerateRandomBytesResponse.data can be performed by computing the CRC32C checksum of GenerateRandomBytesResponse.data and comparing your results to this field. Discard the response in case of non-matching checksum values, and perform a limited number of retries. A persistent mismatch may indicate an issue in your computation of the CRC32C checksum. Note: This field is defined as int64 for reasons of compatibility across different languages. However, it is a non-negative integer, which will never exceed 2^32-1, and can be safely downconverted to uint32 in languages that support this type.
+}
+get(name, x__xgafv=None)
Gets information about a location. diff --git a/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.cryptoKeyVersions.html b/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.cryptoKeyVersions.html index 47f713198d7..14a8c41a525 100644 --- a/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.cryptoKeyVersions.html +++ b/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.cryptoKeyVersions.html @@ -104,6 +104,12 @@Instance Methods
list_next(previous_request, previous_response)Retrieves the next page of results.
++
+macSign(name, body=None, x__xgafv=None)Signs data using a CryptoKeyVersion with CryptoKey.purpose MAC, producing a tag that can be verified by another source with the same key.
++
+macVerify(name, body=None, x__xgafv=None)Verifies MAC tag using a CryptoKeyVersion with CryptoKey.purpose MAC, and returns a response that indicates whether or not the verification was successful.
patch(name, body=None, updateMask=None, x__xgafv=None)Update a CryptoKeyVersion's metadata. state may be changed between ENABLED and DISABLED using this method. See DestroyCryptoKeyVersion and RestoreCryptoKeyVersion to move between other states.
@@ -517,6 +523,71 @@Method Details
macSign(name, body=None, x__xgafv=None)
+ Signs data using a CryptoKeyVersion with CryptoKey.purpose MAC, producing a tag that can be verified by another source with the same key.
+
+Args:
+ name: string, Required. The resource name of the CryptoKeyVersion to use for signing. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for KeyManagementService.MacSign.
+ "data": "A String", # Required. The data to sign. The MAC tag is computed over this data field based on the specific algorithm.
+ "dataCrc32c": "A String", # Optional. An optional CRC32C checksum of the MacSignRequest.data. If specified, KeyManagementService will verify the integrity of the received MacSignRequest.data using this checksum. KeyManagementService will report an error if the checksum verification fails. If you receive a checksum error, your client should verify that CRC32C(MacSignRequest.data) is equal to MacSignRequest.data_crc32c, and if so, perform a limited number of retries. A persistent mismatch may indicate an issue in your computation of the CRC32C checksum. Note: This field is defined as int64 for reasons of compatibility across different languages. However, it is a non-negative integer, which will never exceed 2^32-1, and can be safely downconverted to uint32 in languages that support this type.
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response message for KeyManagementService.MacSign.
+ "mac": "A String", # The created signature.
+ "macCrc32c": "A String", # Integrity verification field. A CRC32C checksum of the returned MacSignResponse.mac. An integrity check of MacSignResponse.mac can be performed by computing the CRC32C checksum of MacSignResponse.mac and comparing your results to this field. Discard the response in case of non-matching checksum values, and perform a limited number of retries. A persistent mismatch may indicate an issue in your computation of the CRC32C checksum. Note: This field is defined as int64 for reasons of compatibility across different languages. However, it is a non-negative integer, which will never exceed 2^32-1, and can be safely downconverted to uint32 in languages that support this type.
+ "name": "A String", # The resource name of the CryptoKeyVersion used for signing. Check this field to verify that the intended resource was used for signing.
+ "protectionLevel": "A String", # The ProtectionLevel of the CryptoKeyVersion used for signing.
+ "verifiedDataCrc32c": True or False, # Integrity verification field. A flag indicating whether MacSignRequest.data_crc32c was received by KeyManagementService and used for the integrity verification of the data. A false value of this field indicates either that MacSignRequest.data_crc32c was left unset or that it was not delivered to KeyManagementService. If you've set MacSignRequest.data_crc32c but this field is still false, discard the response and perform a limited number of retries.
+}
+macVerify(name, body=None, x__xgafv=None)
+ Verifies MAC tag using a CryptoKeyVersion with CryptoKey.purpose MAC, and returns a response that indicates whether or not the verification was successful.
+
+Args:
+ name: string, Required. The resource name of the CryptoKeyVersion to use for verification. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for KeyManagementService.MacVerify.
+ "data": "A String", # Required. The data used previously as a MacSignRequest.data to generate the MAC tag.
+ "dataCrc32c": "A String", # Optional. An optional CRC32C checksum of the MacVerifyRequest.data. If specified, KeyManagementService will verify the integrity of the received MacVerifyRequest.data using this checksum. KeyManagementService will report an error if the checksum verification fails. If you receive a checksum error, your client should verify that CRC32C(MacVerifyRequest.data) is equal to MacVerifyRequest.data_crc32c, and if so, perform a limited number of retries. A persistent mismatch may indicate an issue in your computation of the CRC32C checksum. Note: This field is defined as int64 for reasons of compatibility across different languages. However, it is a non-negative integer, which will never exceed 2^32-1, and can be safely downconverted to uint32 in languages that support this type.
+ "mac": "A String", # Required. The signature to verify.
+ "macCrc32c": "A String", # Optional. An optional CRC32C checksum of the MacVerifyRequest.mac. If specified, KeyManagementService will verify the integrity of the received MacVerifyRequest.mac using this checksum. KeyManagementService will report an error if the checksum verification fails. If you receive a checksum error, your client should verify that CRC32C(MacVerifyRequest.tag) is equal to MacVerifyRequest.mac_crc32c, and if so, perform a limited number of retries. A persistent mismatch may indicate an issue in your computation of the CRC32C checksum. Note: This field is defined as int64 for reasons of compatibility across different languages. However, it is a non-negative integer, which will never exceed 2^32-1, and can be safely downconverted to uint32 in languages that support this type.
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response message for KeyManagementService.MacVerify.
+ "name": "A String", # The resource name of the CryptoKeyVersion used for verification. Check this field to verify that the intended resource was used for verification.
+ "protectionLevel": "A String", # The ProtectionLevel of the CryptoKeyVersion used for verification.
+ "success": True or False, # This field indicates whether or not the verification operation for MacVerifyRequest.mac over MacVerifyRequest.data was successful.
+ "verifiedDataCrc32c": True or False, # Integrity verification field. A flag indicating whether MacVerifyRequest.data_crc32c was received by KeyManagementService and used for the integrity verification of the data. A false value of this field indicates either that MacVerifyRequest.data_crc32c was left unset or that it was not delivered to KeyManagementService. If you've set MacVerifyRequest.data_crc32c but this field is still false, discard the response and perform a limited number of retries.
+ "verifiedMacCrc32c": True or False, # Integrity verification field. A flag indicating whether MacVerifyRequest.mac_crc32c was received by KeyManagementService and used for the integrity verification of the data. A false value of this field indicates either that MacVerifyRequest.mac_crc32c was left unset or that it was not delivered to KeyManagementService. If you've set MacVerifyRequest.mac_crc32c but this field is still false, discard the response and perform a limited number of retries.
+ "verifiedSuccessIntegrity": True or False, # Integrity verification field. This value is used for the integrity verification of [MacVerifyResponse.success]. If the value of this field contradicts the value of [MacVerifyResponse.success], discard the response and perform a limited number of retries.
+}
+patch(name, body=None, updateMask=None, x__xgafv=None)
Update a CryptoKeyVersion's metadata. state may be changed between ENABLED and DISABLED using this method. See DestroyCryptoKeyVersion and RestoreCryptoKeyVersion to move between other states. diff --git a/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.html b/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.html index 22e5f146e92..89cc4209819 100644 --- a/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.html +++ b/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.html @@ -132,6 +132,8 @@Method Details
{ # A CryptoKey represents a logical key that can be used for cryptographic operations. A CryptoKey is made up of zero or more versions, which represent the actual key material used in cryptographic operations. "createTime": "A String", # Output only. The time at which this CryptoKey was created. + "destroyScheduledDuration": "A String", # Immutable. The period of time that versions of this key spend in the DESTROY_SCHEDULED state before transitioning to DESTROYED. If not specified at creation time, the default duration is 24 hours. + "importOnly": True or False, # Immutable. Whether this key may contain imported versions only. "labels": { # Labels with user-defined metadata. For more information, see [Labeling Keys](https://cloud.google.com/kms/docs/labeling-keys). "a_key": "A String", }, @@ -188,6 +190,8 @@Method Details
{ # A CryptoKey represents a logical key that can be used for cryptographic operations. A CryptoKey is made up of zero or more versions, which represent the actual key material used in cryptographic operations. "createTime": "A String", # Output only. The time at which this CryptoKey was created. + "destroyScheduledDuration": "A String", # Immutable. The period of time that versions of this key spend in the DESTROY_SCHEDULED state before transitioning to DESTROYED. If not specified at creation time, the default duration is 24 hours. + "importOnly": True or False, # Immutable. Whether this key may contain imported versions only. "labels": { # Labels with user-defined metadata. For more information, see [Labeling Keys](https://cloud.google.com/kms/docs/labeling-keys). "a_key": "A String", }, @@ -315,6 +319,8 @@Method Details
{ # A CryptoKey represents a logical key that can be used for cryptographic operations. A CryptoKey is made up of zero or more versions, which represent the actual key material used in cryptographic operations. "createTime": "A String", # Output only. The time at which this CryptoKey was created. + "destroyScheduledDuration": "A String", # Immutable. The period of time that versions of this key spend in the DESTROY_SCHEDULED state before transitioning to DESTROYED. If not specified at creation time, the default duration is 24 hours. + "importOnly": True or False, # Immutable. Whether this key may contain imported versions only. "labels": { # Labels with user-defined metadata. For more information, see [Labeling Keys](https://cloud.google.com/kms/docs/labeling-keys). "a_key": "A String", }, @@ -434,6 +440,8 @@Method Details
"cryptoKeys": [ # The list of CryptoKeys. { # A CryptoKey represents a logical key that can be used for cryptographic operations. A CryptoKey is made up of zero or more versions, which represent the actual key material used in cryptographic operations. "createTime": "A String", # Output only. The time at which this CryptoKey was created. + "destroyScheduledDuration": "A String", # Immutable. The period of time that versions of this key spend in the DESTROY_SCHEDULED state before transitioning to DESTROYED. If not specified at creation time, the default duration is 24 hours. + "importOnly": True or False, # Immutable. Whether this key may contain imported versions only. "labels": { # Labels with user-defined metadata. For more information, see [Labeling Keys](https://cloud.google.com/kms/docs/labeling-keys). "a_key": "A String", }, @@ -508,6 +516,8 @@Method Details
{ # A CryptoKey represents a logical key that can be used for cryptographic operations. A CryptoKey is made up of zero or more versions, which represent the actual key material used in cryptographic operations. "createTime": "A String", # Output only. The time at which this CryptoKey was created. + "destroyScheduledDuration": "A String", # Immutable. The period of time that versions of this key spend in the DESTROY_SCHEDULED state before transitioning to DESTROYED. If not specified at creation time, the default duration is 24 hours. + "importOnly": True or False, # Immutable. Whether this key may contain imported versions only. "labels": { # Labels with user-defined metadata. For more information, see [Labeling Keys](https://cloud.google.com/kms/docs/labeling-keys). "a_key": "A String", }, @@ -563,6 +573,8 @@Method Details
{ # A CryptoKey represents a logical key that can be used for cryptographic operations. A CryptoKey is made up of zero or more versions, which represent the actual key material used in cryptographic operations. "createTime": "A String", # Output only. The time at which this CryptoKey was created. + "destroyScheduledDuration": "A String", # Immutable. The period of time that versions of this key spend in the DESTROY_SCHEDULED state before transitioning to DESTROYED. If not specified at creation time, the default duration is 24 hours. + "importOnly": True or False, # Immutable. Whether this key may contain imported versions only. "labels": { # Labels with user-defined metadata. For more information, see [Labeling Keys](https://cloud.google.com/kms/docs/labeling-keys). "a_key": "A String", }, @@ -746,6 +758,8 @@Method Details
{ # A CryptoKey represents a logical key that can be used for cryptographic operations. A CryptoKey is made up of zero or more versions, which represent the actual key material used in cryptographic operations. "createTime": "A String", # Output only. The time at which this CryptoKey was created. + "destroyScheduledDuration": "A String", # Immutable. The period of time that versions of this key spend in the DESTROY_SCHEDULED state before transitioning to DESTROYED. If not specified at creation time, the default duration is 24 hours. + "importOnly": True or False, # Immutable. Whether this key may contain imported versions only. "labels": { # Labels with user-defined metadata. For more information, see [Labeling Keys](https://cloud.google.com/kms/docs/labeling-keys). "a_key": "A String", }, diff --git a/docs/dyn/composer_v1beta1.projects.locations.environments.html b/docs/dyn/composer_v1beta1.projects.locations.environments.html index 769a7f7e0e0..b47a47fa107 100644 --- a/docs/dyn/composer_v1beta1.projects.locations.environments.html +++ b/docs/dyn/composer_v1beta1.projects.locations.environments.html @@ -162,10 +162,10 @@Method Details
"config": { # Configuration information for an environment. # Configuration parameters for this environment. "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. - "databaseConfig": { # The configuration of Cloud SQL instance that is used by the Apache Airflow software. # Optional. The configuration settings for Cloud SQL instance used internally by Apache Airflow software. + "databaseConfig": { # The configuration of Cloud SQL instance that is used by the Apache Airflow software. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for Cloud SQL instance used internally by Apache Airflow software. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Cloud SQL machine type used by Airflow database. It has to be one of: db-n1-standard-2, db-n1-standard-4, db-n1-standard-8 or db-n1-standard-16. If not specified, db-n1-standard-2 will be used. }, - "encryptionConfig": { # The encryption options for the Cloud Composer environment and its dependencies. # Optional. The encryption options for the Cloud Composer environment and its dependencies. Cannot be updated. + "encryptionConfig": { # The encryption options for the Cloud Composer environment and its dependencies. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The encryption options for the Cloud Composer environment and its dependencies. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "kmsKeyName": "A String", # Optional. Customer-managed Encryption Key available through Google's Key Management Service. Cannot be updated. If not specified, Google-managed key will be used. }, "environmentSize": "A String", # Optional. The size of the Cloud Composer environment. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -176,40 +176,42 @@Method Details
"startTime": "A String", # Required. Start time of the first recurrence of the maintenance window. }, "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. - "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 20GB. If unspecified, defaults to 100GB. Cannot be updated. + "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 20GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "enableIpMasqAgent": True or False, # Optional. Deploys 'ip-masq-agent' daemon set in the GKE cluster and defines nonMasqueradeCIDRs equals to pod IP range so IP masquerading is used for all destination addresses, except between pods traffic. See: https://cloud.google.com/kubernetes-engine/docs/how-to/ip-masquerade-agent "ipAllocationPolicy": { # Configuration for controlling how IPs are allocated in the GKE cluster. # Optional. The IPAllocationPolicy fields for the GKE cluster. - "clusterIpv4CidrBlock": "A String", # Optional. The IP address range used to allocate IP addresses to pods in the cluster. This field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. - "clusterSecondaryRangeName": "A String", # Optional. The name of the cluster's secondary range used to allocate IP addresses to pods. Specify either `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. - "servicesIpv4CidrBlock": "A String", # Optional. The IP address range of the services IP addresses in this cluster. This field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. - "servicesSecondaryRangeName": "A String", # Optional. The name of the services' secondary range used to allocate IP addresses to the cluster. Specify either `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. - "useIpAliases": True or False, # Optional. Whether or not to enable Alias IPs in the GKE cluster. If `true`, a VPC-native cluster is created. + "clusterIpv4CidrBlock": "A String", # Optional. The IP address range used to allocate IP addresses to pods in the cluster. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. + "clusterSecondaryRangeName": "A String", # Optional. The name of the cluster's secondary range used to allocate IP addresses to pods. Specify either `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. + "servicesIpv4CidrBlock": "A String", # Optional. The IP address range of the services IP addresses in this cluster. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. + "servicesSecondaryRangeName": "A String", # Optional. The name of the services' secondary range used to allocate IP addresses to the cluster. Specify either `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. + "useIpAliases": True or False, # Optional. Whether or not to enable Alias IPs in the GKE cluster. If `true`, a VPC-native cluster is created. This field is only supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. Environments in newer versions always use VPC-native GKE clusters. }, - "location": "A String", # Optional. The Compute Engine [zone](/compute/docs/regions-zones) in which to deploy the VMs used to run the Apache Airflow software, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}". This `location` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.machineType` are specified, `nodeConfig.machineType` must belong to this `location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If only one field (`location` or `nodeConfig.machineType`) is specified, the location information from the specified field will be propagated to the unspecified field. - "machineType": "A String", # Optional. The Compute Engine [machine type](/compute/docs/machine-types) used for cluster instances, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}/machineTypes/{machineTypeId}". The `machineType` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.location` are specified, this `machineType` must belong to the `nodeConfig.location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If exactly one of this field and `nodeConfig.location` is specified, the location information from the specified field will be propagated to the unspecified field. The `machineTypeId` must not be a [shared-core machine type](/compute/docs/machine-types#sharedcore). If this field is unspecified, the `machineTypeId` defaults to "n1-standard-1". - "maxPodsPerNode": 42, # Optional. The maximum number of pods per node in the Cloud Composer GKE cluster. The value must be between 8 and 110 and it can be set only if the environment is VPC-native. The default value is 32. Values of this field will be propagated both to the `default-pool` node pool of the newly created GKE cluster, and to the default "Maximum Pods per Node" value which is used for newly created node pools if their value is not explicitly set during node pool creation. For more information, see [Optimizing IP address allocation] (https://cloud.google.com/kubernetes-engine/docs/how-to/flexible-pod-cidr). Cannot be updated. + "location": "A String", # Optional. The Compute Engine [zone](/compute/docs/regions-zones) in which to deploy the VMs used to run the Apache Airflow software, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}". This `location` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.machineType` are specified, `nodeConfig.machineType` must belong to this `location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If only one field (`location` or `nodeConfig.machineType`) is specified, the location information from the specified field will be propagated to the unspecified field. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "machineType": "A String", # Optional. The Compute Engine [machine type](/compute/docs/machine-types) used for cluster instances, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}/machineTypes/{machineTypeId}". The `machineType` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.location` are specified, this `machineType` must belong to the `nodeConfig.location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If exactly one of this field and `nodeConfig.location` is specified, the location information from the specified field will be propagated to the unspecified field. The `machineTypeId` must not be a [shared-core machine type](/compute/docs/machine-types#sharedcore). If this field is unspecified, the `machineTypeId` defaults to "n1-standard-1". This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "maxPodsPerNode": 42, # Optional. The maximum number of pods per node in the Cloud Composer GKE cluster. The value must be between 8 and 110 and it can be set only if the environment is VPC-native. The default value is 32. Values of this field will be propagated both to the `default-pool` node pool of the newly created GKE cluster, and to the default "Maximum Pods per Node" value which is used for newly created node pools if their value is not explicitly set during node pool creation. For more information, see [Optimizing IP address allocation] (https://cloud.google.com/kubernetes-engine/docs/how-to/flexible-pod-cidr). Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "network": "A String", # Optional. The Compute Engine network to be used for machine communications, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/global/networks/{networkId}". If unspecified, the default network in the environment's project is used. If a [Custom Subnet Network](/vpc/docs/vpc#vpc_networks_and_subnets) is provided, `nodeConfig.subnetwork` must also be provided. For [Shared VPC](/vpc/docs/shared-vpc) subnetwork requirements, see `nodeConfig.subnetwork`. - "oauthScopes": [ # Optional. The set of Google API scopes to be made available on all node VMs. If `oauth_scopes` is empty, defaults to ["https://www.googleapis.com/auth/cloud-platform"]. Cannot be updated. + "oauthScopes": [ # Optional. The set of Google API scopes to be made available on all node VMs. If `oauth_scopes` is empty, defaults to ["https://www.googleapis.com/auth/cloud-platform"]. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "A String", ], "serviceAccount": "A String", # Optional. The Google Cloud Platform Service Account to be used by the workloads. If a service account is not specified, the "default" Compute Engine service account is used. Cannot be updated. "subnetwork": "A String", # Optional. The Compute Engine subnetwork to be used for machine communications, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/regions/{regionId}/subnetworks/{subnetworkId}" If a subnetwork is provided, `nodeConfig.network` must also be provided, and the subnetwork must belong to the enclosing environment's project and location. - "tags": [ # Optional. The list of instance tags applied to all node VMs. Tags are used to identify valid sources or targets for network firewalls. Each tag within the list must comply with [RFC1035](https://www.ietf.org/rfc/rfc1035.txt). Cannot be updated. + "tags": [ # Optional. The list of instance tags applied to all node VMs. Tags are used to identify valid sources or targets for network firewalls. Each tag within the list must comply with [RFC1035](https://www.ietf.org/rfc/rfc1035.txt). Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "A String", ], }, - "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. + "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudSqlIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range in tenant project will be reserved for Cloud SQL. Needs to be disjoint from web_server_ipv4_cidr_block - "enablePrivateEnvironment": True or False, # Optional. If `true`, a Private IP Cloud Composer environment is created. If this field is set to true, `IPAllocationPolicy.use_ip_aliases` must be set to true . + "enablePrivateEnvironment": True or False, # Optional. If `true`, a Private IP Cloud Composer environment is created. If this field is set to true, `IPAllocationPolicy.use_ip_aliases` must be set to true for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "enablePrivatelyUsedPublicIps": True or False, # Optional. When enabled, IPs from public (non-RFC1918) ranges can be used for `IPAllocationPolicy.cluster_ipv4_cidr_block` and `IPAllocationPolicy.service_ipv4_cidr_block`. "privateClusterConfig": { # Configuration options for the private GKE cluster in a Cloud Composer environment. # Optional. Configuration for the private GKE cluster for a Private IP Cloud Composer environment. "enablePrivateEndpoint": True or False, # Optional. If `true`, access to the public endpoint of the GKE cluster is denied. "masterIpv4CidrBlock": "A String", # Optional. The CIDR block from which IPv4 range for GKE master will be reserved. If left blank, the default value of '172.16.0.0/23' is used. "masterIpv4ReservedRange": "A String", # Output only. The IP range in CIDR notation to use for the hosted master network. This range is used for assigning internal IP addresses to the cluster master or set of masters and to the internal load balancer virtual IP. This range must not overlap with any other ranges in use within the cluster's network. }, - "webServerIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for web server will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. - "webServerIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's App Engine VMs. + "webServerIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for web server will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's App Engine VMs. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. }, "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. @@ -222,13 +224,13 @@Method Details
"pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, - "pythonVersion": "A String", # Optional. The major version of Python used to run the Apache Airflow scheduler, worker, and webserver processes. Can be set to '2' or '3'. If not specified, the default is '3'. Cannot be updated. - "schedulerCount": 42, # Optional. The number of schedulers for Airflow. + "pythonVersion": "A String", # Optional. The major version of Python used to run the Apache Airflow scheduler, worker, and webserver processes. Can be set to '2' or '3'. If not specified, the default is '3'. Cannot be updated. This field is only supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. Environments in newer versions always use Python major version 3. + "schedulerCount": 42, # Optional. The number of schedulers for Airflow. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-2.*.*. }, - "webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. # Optional. The configuration settings for the Airflow web server App Engine instance. + "webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for the Airflow web server App Engine instance. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. @@ -349,10 +351,10 @@Method Details
"config": { # Configuration information for an environment. # Configuration parameters for this environment. "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. - "databaseConfig": { # The configuration of Cloud SQL instance that is used by the Apache Airflow software. # Optional. The configuration settings for Cloud SQL instance used internally by Apache Airflow software. + "databaseConfig": { # The configuration of Cloud SQL instance that is used by the Apache Airflow software. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for Cloud SQL instance used internally by Apache Airflow software. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Cloud SQL machine type used by Airflow database. It has to be one of: db-n1-standard-2, db-n1-standard-4, db-n1-standard-8 or db-n1-standard-16. If not specified, db-n1-standard-2 will be used. }, - "encryptionConfig": { # The encryption options for the Cloud Composer environment and its dependencies. # Optional. The encryption options for the Cloud Composer environment and its dependencies. Cannot be updated. + "encryptionConfig": { # The encryption options for the Cloud Composer environment and its dependencies. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The encryption options for the Cloud Composer environment and its dependencies. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "kmsKeyName": "A String", # Optional. Customer-managed Encryption Key available through Google's Key Management Service. Cannot be updated. If not specified, Google-managed key will be used. }, "environmentSize": "A String", # Optional. The size of the Cloud Composer environment. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -363,40 +365,42 @@Method Details
"startTime": "A String", # Required. Start time of the first recurrence of the maintenance window. }, "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. - "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 20GB. If unspecified, defaults to 100GB. Cannot be updated. + "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 20GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "enableIpMasqAgent": True or False, # Optional. Deploys 'ip-masq-agent' daemon set in the GKE cluster and defines nonMasqueradeCIDRs equals to pod IP range so IP masquerading is used for all destination addresses, except between pods traffic. See: https://cloud.google.com/kubernetes-engine/docs/how-to/ip-masquerade-agent "ipAllocationPolicy": { # Configuration for controlling how IPs are allocated in the GKE cluster. # Optional. The IPAllocationPolicy fields for the GKE cluster. - "clusterIpv4CidrBlock": "A String", # Optional. The IP address range used to allocate IP addresses to pods in the cluster. This field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. - "clusterSecondaryRangeName": "A String", # Optional. The name of the cluster's secondary range used to allocate IP addresses to pods. Specify either `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. - "servicesIpv4CidrBlock": "A String", # Optional. The IP address range of the services IP addresses in this cluster. This field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. - "servicesSecondaryRangeName": "A String", # Optional. The name of the services' secondary range used to allocate IP addresses to the cluster. Specify either `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. - "useIpAliases": True or False, # Optional. Whether or not to enable Alias IPs in the GKE cluster. If `true`, a VPC-native cluster is created. + "clusterIpv4CidrBlock": "A String", # Optional. The IP address range used to allocate IP addresses to pods in the cluster. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. + "clusterSecondaryRangeName": "A String", # Optional. The name of the cluster's secondary range used to allocate IP addresses to pods. Specify either `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. + "servicesIpv4CidrBlock": "A String", # Optional. The IP address range of the services IP addresses in this cluster. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. + "servicesSecondaryRangeName": "A String", # Optional. The name of the services' secondary range used to allocate IP addresses to the cluster. Specify either `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. + "useIpAliases": True or False, # Optional. Whether or not to enable Alias IPs in the GKE cluster. If `true`, a VPC-native cluster is created. This field is only supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. Environments in newer versions always use VPC-native GKE clusters. }, - "location": "A String", # Optional. The Compute Engine [zone](/compute/docs/regions-zones) in which to deploy the VMs used to run the Apache Airflow software, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}". This `location` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.machineType` are specified, `nodeConfig.machineType` must belong to this `location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If only one field (`location` or `nodeConfig.machineType`) is specified, the location information from the specified field will be propagated to the unspecified field. - "machineType": "A String", # Optional. The Compute Engine [machine type](/compute/docs/machine-types) used for cluster instances, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}/machineTypes/{machineTypeId}". The `machineType` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.location` are specified, this `machineType` must belong to the `nodeConfig.location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If exactly one of this field and `nodeConfig.location` is specified, the location information from the specified field will be propagated to the unspecified field. The `machineTypeId` must not be a [shared-core machine type](/compute/docs/machine-types#sharedcore). If this field is unspecified, the `machineTypeId` defaults to "n1-standard-1". - "maxPodsPerNode": 42, # Optional. The maximum number of pods per node in the Cloud Composer GKE cluster. The value must be between 8 and 110 and it can be set only if the environment is VPC-native. The default value is 32. Values of this field will be propagated both to the `default-pool` node pool of the newly created GKE cluster, and to the default "Maximum Pods per Node" value which is used for newly created node pools if their value is not explicitly set during node pool creation. For more information, see [Optimizing IP address allocation] (https://cloud.google.com/kubernetes-engine/docs/how-to/flexible-pod-cidr). Cannot be updated. + "location": "A String", # Optional. The Compute Engine [zone](/compute/docs/regions-zones) in which to deploy the VMs used to run the Apache Airflow software, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}". This `location` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.machineType` are specified, `nodeConfig.machineType` must belong to this `location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If only one field (`location` or `nodeConfig.machineType`) is specified, the location information from the specified field will be propagated to the unspecified field. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "machineType": "A String", # Optional. The Compute Engine [machine type](/compute/docs/machine-types) used for cluster instances, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}/machineTypes/{machineTypeId}". The `machineType` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.location` are specified, this `machineType` must belong to the `nodeConfig.location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If exactly one of this field and `nodeConfig.location` is specified, the location information from the specified field will be propagated to the unspecified field. The `machineTypeId` must not be a [shared-core machine type](/compute/docs/machine-types#sharedcore). If this field is unspecified, the `machineTypeId` defaults to "n1-standard-1". This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "maxPodsPerNode": 42, # Optional. The maximum number of pods per node in the Cloud Composer GKE cluster. The value must be between 8 and 110 and it can be set only if the environment is VPC-native. The default value is 32. Values of this field will be propagated both to the `default-pool` node pool of the newly created GKE cluster, and to the default "Maximum Pods per Node" value which is used for newly created node pools if their value is not explicitly set during node pool creation. For more information, see [Optimizing IP address allocation] (https://cloud.google.com/kubernetes-engine/docs/how-to/flexible-pod-cidr). Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "network": "A String", # Optional. The Compute Engine network to be used for machine communications, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/global/networks/{networkId}". If unspecified, the default network in the environment's project is used. If a [Custom Subnet Network](/vpc/docs/vpc#vpc_networks_and_subnets) is provided, `nodeConfig.subnetwork` must also be provided. For [Shared VPC](/vpc/docs/shared-vpc) subnetwork requirements, see `nodeConfig.subnetwork`. - "oauthScopes": [ # Optional. The set of Google API scopes to be made available on all node VMs. If `oauth_scopes` is empty, defaults to ["https://www.googleapis.com/auth/cloud-platform"]. Cannot be updated. + "oauthScopes": [ # Optional. The set of Google API scopes to be made available on all node VMs. If `oauth_scopes` is empty, defaults to ["https://www.googleapis.com/auth/cloud-platform"]. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "A String", ], "serviceAccount": "A String", # Optional. The Google Cloud Platform Service Account to be used by the workloads. If a service account is not specified, the "default" Compute Engine service account is used. Cannot be updated. "subnetwork": "A String", # Optional. The Compute Engine subnetwork to be used for machine communications, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/regions/{regionId}/subnetworks/{subnetworkId}" If a subnetwork is provided, `nodeConfig.network` must also be provided, and the subnetwork must belong to the enclosing environment's project and location. - "tags": [ # Optional. The list of instance tags applied to all node VMs. Tags are used to identify valid sources or targets for network firewalls. Each tag within the list must comply with [RFC1035](https://www.ietf.org/rfc/rfc1035.txt). Cannot be updated. + "tags": [ # Optional. The list of instance tags applied to all node VMs. Tags are used to identify valid sources or targets for network firewalls. Each tag within the list must comply with [RFC1035](https://www.ietf.org/rfc/rfc1035.txt). Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "A String", ], }, - "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. + "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudSqlIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range in tenant project will be reserved for Cloud SQL. Needs to be disjoint from web_server_ipv4_cidr_block - "enablePrivateEnvironment": True or False, # Optional. If `true`, a Private IP Cloud Composer environment is created. If this field is set to true, `IPAllocationPolicy.use_ip_aliases` must be set to true . + "enablePrivateEnvironment": True or False, # Optional. If `true`, a Private IP Cloud Composer environment is created. If this field is set to true, `IPAllocationPolicy.use_ip_aliases` must be set to true for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "enablePrivatelyUsedPublicIps": True or False, # Optional. When enabled, IPs from public (non-RFC1918) ranges can be used for `IPAllocationPolicy.cluster_ipv4_cidr_block` and `IPAllocationPolicy.service_ipv4_cidr_block`. "privateClusterConfig": { # Configuration options for the private GKE cluster in a Cloud Composer environment. # Optional. Configuration for the private GKE cluster for a Private IP Cloud Composer environment. "enablePrivateEndpoint": True or False, # Optional. If `true`, access to the public endpoint of the GKE cluster is denied. "masterIpv4CidrBlock": "A String", # Optional. The CIDR block from which IPv4 range for GKE master will be reserved. If left blank, the default value of '172.16.0.0/23' is used. "masterIpv4ReservedRange": "A String", # Output only. The IP range in CIDR notation to use for the hosted master network. This range is used for assigning internal IP addresses to the cluster master or set of masters and to the internal load balancer virtual IP. This range must not overlap with any other ranges in use within the cluster's network. }, - "webServerIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for web server will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. - "webServerIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's App Engine VMs. + "webServerIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for web server will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's App Engine VMs. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. }, "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. @@ -409,13 +413,13 @@Method Details
"pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, - "pythonVersion": "A String", # Optional. The major version of Python used to run the Apache Airflow scheduler, worker, and webserver processes. Can be set to '2' or '3'. If not specified, the default is '3'. Cannot be updated. - "schedulerCount": 42, # Optional. The number of schedulers for Airflow. + "pythonVersion": "A String", # Optional. The major version of Python used to run the Apache Airflow scheduler, worker, and webserver processes. Can be set to '2' or '3'. If not specified, the default is '3'. Cannot be updated. This field is only supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. Environments in newer versions always use Python major version 3. + "schedulerCount": 42, # Optional. The number of schedulers for Airflow. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-2.*.*. }, - "webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. # Optional. The configuration settings for the Airflow web server App Engine instance. + "webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for the Airflow web server App Engine instance. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. @@ -477,10 +481,10 @@Method Details
"config": { # Configuration information for an environment. # Configuration parameters for this environment. "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. - "databaseConfig": { # The configuration of Cloud SQL instance that is used by the Apache Airflow software. # Optional. The configuration settings for Cloud SQL instance used internally by Apache Airflow software. + "databaseConfig": { # The configuration of Cloud SQL instance that is used by the Apache Airflow software. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for Cloud SQL instance used internally by Apache Airflow software. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Cloud SQL machine type used by Airflow database. It has to be one of: db-n1-standard-2, db-n1-standard-4, db-n1-standard-8 or db-n1-standard-16. If not specified, db-n1-standard-2 will be used. }, - "encryptionConfig": { # The encryption options for the Cloud Composer environment and its dependencies. # Optional. The encryption options for the Cloud Composer environment and its dependencies. Cannot be updated. + "encryptionConfig": { # The encryption options for the Cloud Composer environment and its dependencies. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The encryption options for the Cloud Composer environment and its dependencies. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "kmsKeyName": "A String", # Optional. Customer-managed Encryption Key available through Google's Key Management Service. Cannot be updated. If not specified, Google-managed key will be used. }, "environmentSize": "A String", # Optional. The size of the Cloud Composer environment. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -491,40 +495,42 @@Method Details
"startTime": "A String", # Required. Start time of the first recurrence of the maintenance window. }, "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. - "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 20GB. If unspecified, defaults to 100GB. Cannot be updated. + "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 20GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "enableIpMasqAgent": True or False, # Optional. Deploys 'ip-masq-agent' daemon set in the GKE cluster and defines nonMasqueradeCIDRs equals to pod IP range so IP masquerading is used for all destination addresses, except between pods traffic. See: https://cloud.google.com/kubernetes-engine/docs/how-to/ip-masquerade-agent "ipAllocationPolicy": { # Configuration for controlling how IPs are allocated in the GKE cluster. # Optional. The IPAllocationPolicy fields for the GKE cluster. - "clusterIpv4CidrBlock": "A String", # Optional. The IP address range used to allocate IP addresses to pods in the cluster. This field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. - "clusterSecondaryRangeName": "A String", # Optional. The name of the cluster's secondary range used to allocate IP addresses to pods. Specify either `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. - "servicesIpv4CidrBlock": "A String", # Optional. The IP address range of the services IP addresses in this cluster. This field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. - "servicesSecondaryRangeName": "A String", # Optional. The name of the services' secondary range used to allocate IP addresses to the cluster. Specify either `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. - "useIpAliases": True or False, # Optional. Whether or not to enable Alias IPs in the GKE cluster. If `true`, a VPC-native cluster is created. + "clusterIpv4CidrBlock": "A String", # Optional. The IP address range used to allocate IP addresses to pods in the cluster. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. + "clusterSecondaryRangeName": "A String", # Optional. The name of the cluster's secondary range used to allocate IP addresses to pods. Specify either `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. + "servicesIpv4CidrBlock": "A String", # Optional. The IP address range of the services IP addresses in this cluster. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. + "servicesSecondaryRangeName": "A String", # Optional. The name of the services' secondary range used to allocate IP addresses to the cluster. Specify either `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. + "useIpAliases": True or False, # Optional. Whether or not to enable Alias IPs in the GKE cluster. If `true`, a VPC-native cluster is created. This field is only supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. Environments in newer versions always use VPC-native GKE clusters. }, - "location": "A String", # Optional. The Compute Engine [zone](/compute/docs/regions-zones) in which to deploy the VMs used to run the Apache Airflow software, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}". This `location` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.machineType` are specified, `nodeConfig.machineType` must belong to this `location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If only one field (`location` or `nodeConfig.machineType`) is specified, the location information from the specified field will be propagated to the unspecified field. - "machineType": "A String", # Optional. The Compute Engine [machine type](/compute/docs/machine-types) used for cluster instances, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}/machineTypes/{machineTypeId}". The `machineType` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.location` are specified, this `machineType` must belong to the `nodeConfig.location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If exactly one of this field and `nodeConfig.location` is specified, the location information from the specified field will be propagated to the unspecified field. The `machineTypeId` must not be a [shared-core machine type](/compute/docs/machine-types#sharedcore). If this field is unspecified, the `machineTypeId` defaults to "n1-standard-1". - "maxPodsPerNode": 42, # Optional. The maximum number of pods per node in the Cloud Composer GKE cluster. The value must be between 8 and 110 and it can be set only if the environment is VPC-native. The default value is 32. Values of this field will be propagated both to the `default-pool` node pool of the newly created GKE cluster, and to the default "Maximum Pods per Node" value which is used for newly created node pools if their value is not explicitly set during node pool creation. For more information, see [Optimizing IP address allocation] (https://cloud.google.com/kubernetes-engine/docs/how-to/flexible-pod-cidr). Cannot be updated. + "location": "A String", # Optional. The Compute Engine [zone](/compute/docs/regions-zones) in which to deploy the VMs used to run the Apache Airflow software, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}". This `location` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.machineType` are specified, `nodeConfig.machineType` must belong to this `location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If only one field (`location` or `nodeConfig.machineType`) is specified, the location information from the specified field will be propagated to the unspecified field. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "machineType": "A String", # Optional. The Compute Engine [machine type](/compute/docs/machine-types) used for cluster instances, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}/machineTypes/{machineTypeId}". The `machineType` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.location` are specified, this `machineType` must belong to the `nodeConfig.location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If exactly one of this field and `nodeConfig.location` is specified, the location information from the specified field will be propagated to the unspecified field. The `machineTypeId` must not be a [shared-core machine type](/compute/docs/machine-types#sharedcore). If this field is unspecified, the `machineTypeId` defaults to "n1-standard-1". This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "maxPodsPerNode": 42, # Optional. The maximum number of pods per node in the Cloud Composer GKE cluster. The value must be between 8 and 110 and it can be set only if the environment is VPC-native. The default value is 32. Values of this field will be propagated both to the `default-pool` node pool of the newly created GKE cluster, and to the default "Maximum Pods per Node" value which is used for newly created node pools if their value is not explicitly set during node pool creation. For more information, see [Optimizing IP address allocation] (https://cloud.google.com/kubernetes-engine/docs/how-to/flexible-pod-cidr). Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "network": "A String", # Optional. The Compute Engine network to be used for machine communications, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/global/networks/{networkId}". If unspecified, the default network in the environment's project is used. If a [Custom Subnet Network](/vpc/docs/vpc#vpc_networks_and_subnets) is provided, `nodeConfig.subnetwork` must also be provided. For [Shared VPC](/vpc/docs/shared-vpc) subnetwork requirements, see `nodeConfig.subnetwork`. - "oauthScopes": [ # Optional. The set of Google API scopes to be made available on all node VMs. If `oauth_scopes` is empty, defaults to ["https://www.googleapis.com/auth/cloud-platform"]. Cannot be updated. + "oauthScopes": [ # Optional. The set of Google API scopes to be made available on all node VMs. If `oauth_scopes` is empty, defaults to ["https://www.googleapis.com/auth/cloud-platform"]. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "A String", ], "serviceAccount": "A String", # Optional. The Google Cloud Platform Service Account to be used by the workloads. If a service account is not specified, the "default" Compute Engine service account is used. Cannot be updated. "subnetwork": "A String", # Optional. The Compute Engine subnetwork to be used for machine communications, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/regions/{regionId}/subnetworks/{subnetworkId}" If a subnetwork is provided, `nodeConfig.network` must also be provided, and the subnetwork must belong to the enclosing environment's project and location. - "tags": [ # Optional. The list of instance tags applied to all node VMs. Tags are used to identify valid sources or targets for network firewalls. Each tag within the list must comply with [RFC1035](https://www.ietf.org/rfc/rfc1035.txt). Cannot be updated. + "tags": [ # Optional. The list of instance tags applied to all node VMs. Tags are used to identify valid sources or targets for network firewalls. Each tag within the list must comply with [RFC1035](https://www.ietf.org/rfc/rfc1035.txt). Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "A String", ], }, - "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. + "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudSqlIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range in tenant project will be reserved for Cloud SQL. Needs to be disjoint from web_server_ipv4_cidr_block - "enablePrivateEnvironment": True or False, # Optional. If `true`, a Private IP Cloud Composer environment is created. If this field is set to true, `IPAllocationPolicy.use_ip_aliases` must be set to true . + "enablePrivateEnvironment": True or False, # Optional. If `true`, a Private IP Cloud Composer environment is created. If this field is set to true, `IPAllocationPolicy.use_ip_aliases` must be set to true for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "enablePrivatelyUsedPublicIps": True or False, # Optional. When enabled, IPs from public (non-RFC1918) ranges can be used for `IPAllocationPolicy.cluster_ipv4_cidr_block` and `IPAllocationPolicy.service_ipv4_cidr_block`. "privateClusterConfig": { # Configuration options for the private GKE cluster in a Cloud Composer environment. # Optional. Configuration for the private GKE cluster for a Private IP Cloud Composer environment. "enablePrivateEndpoint": True or False, # Optional. If `true`, access to the public endpoint of the GKE cluster is denied. "masterIpv4CidrBlock": "A String", # Optional. The CIDR block from which IPv4 range for GKE master will be reserved. If left blank, the default value of '172.16.0.0/23' is used. "masterIpv4ReservedRange": "A String", # Output only. The IP range in CIDR notation to use for the hosted master network. This range is used for assigning internal IP addresses to the cluster master or set of masters and to the internal load balancer virtual IP. This range must not overlap with any other ranges in use within the cluster's network. }, - "webServerIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for web server will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. - "webServerIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's App Engine VMs. + "webServerIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for web server will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's App Engine VMs. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. }, "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. @@ -537,13 +543,13 @@Method Details
"pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, - "pythonVersion": "A String", # Optional. The major version of Python used to run the Apache Airflow scheduler, worker, and webserver processes. Can be set to '2' or '3'. If not specified, the default is '3'. Cannot be updated. - "schedulerCount": 42, # Optional. The number of schedulers for Airflow. + "pythonVersion": "A String", # Optional. The major version of Python used to run the Apache Airflow scheduler, worker, and webserver processes. Can be set to '2' or '3'. If not specified, the default is '3'. Cannot be updated. This field is only supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. Environments in newer versions always use Python major version 3. + "schedulerCount": 42, # Optional. The number of schedulers for Airflow. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-2.*.*. }, - "webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. # Optional. The configuration settings for the Airflow web server App Engine instance. + "webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for the Airflow web server App Engine instance. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. @@ -613,10 +619,10 @@Method Details
"config": { # Configuration information for an environment. # Configuration parameters for this environment. "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. - "databaseConfig": { # The configuration of Cloud SQL instance that is used by the Apache Airflow software. # Optional. The configuration settings for Cloud SQL instance used internally by Apache Airflow software. + "databaseConfig": { # The configuration of Cloud SQL instance that is used by the Apache Airflow software. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for Cloud SQL instance used internally by Apache Airflow software. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Cloud SQL machine type used by Airflow database. It has to be one of: db-n1-standard-2, db-n1-standard-4, db-n1-standard-8 or db-n1-standard-16. If not specified, db-n1-standard-2 will be used. }, - "encryptionConfig": { # The encryption options for the Cloud Composer environment and its dependencies. # Optional. The encryption options for the Cloud Composer environment and its dependencies. Cannot be updated. + "encryptionConfig": { # The encryption options for the Cloud Composer environment and its dependencies. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The encryption options for the Cloud Composer environment and its dependencies. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "kmsKeyName": "A String", # Optional. Customer-managed Encryption Key available through Google's Key Management Service. Cannot be updated. If not specified, Google-managed key will be used. }, "environmentSize": "A String", # Optional. The size of the Cloud Composer environment. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -627,40 +633,42 @@Method Details
"startTime": "A String", # Required. Start time of the first recurrence of the maintenance window. }, "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. - "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 20GB. If unspecified, defaults to 100GB. Cannot be updated. + "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 20GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "enableIpMasqAgent": True or False, # Optional. Deploys 'ip-masq-agent' daemon set in the GKE cluster and defines nonMasqueradeCIDRs equals to pod IP range so IP masquerading is used for all destination addresses, except between pods traffic. See: https://cloud.google.com/kubernetes-engine/docs/how-to/ip-masquerade-agent "ipAllocationPolicy": { # Configuration for controlling how IPs are allocated in the GKE cluster. # Optional. The IPAllocationPolicy fields for the GKE cluster. - "clusterIpv4CidrBlock": "A String", # Optional. The IP address range used to allocate IP addresses to pods in the cluster. This field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. - "clusterSecondaryRangeName": "A String", # Optional. The name of the cluster's secondary range used to allocate IP addresses to pods. Specify either `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. - "servicesIpv4CidrBlock": "A String", # Optional. The IP address range of the services IP addresses in this cluster. This field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. - "servicesSecondaryRangeName": "A String", # Optional. The name of the services' secondary range used to allocate IP addresses to the cluster. Specify either `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. - "useIpAliases": True or False, # Optional. Whether or not to enable Alias IPs in the GKE cluster. If `true`, a VPC-native cluster is created. + "clusterIpv4CidrBlock": "A String", # Optional. The IP address range used to allocate IP addresses to pods in the cluster. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. + "clusterSecondaryRangeName": "A String", # Optional. The name of the cluster's secondary range used to allocate IP addresses to pods. Specify either `cluster_secondary_range_name` or `cluster_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. + "servicesIpv4CidrBlock": "A String", # Optional. The IP address range of the services IP addresses in this cluster. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. Set to blank to have GKE choose a range with the default size. Set to /netmask (e.g. `/14`) to have GKE choose a range with a specific netmask. Set to a [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range to use. Specify `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. + "servicesSecondaryRangeName": "A String", # Optional. The name of the services' secondary range used to allocate IP addresses to the cluster. Specify either `services_secondary_range_name` or `services_ipv4_cidr_block` but not both. This field is applicable only when `use_ip_aliases` is true. For Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*, this field is applicable only when `use_ip_aliases` is true. + "useIpAliases": True or False, # Optional. Whether or not to enable Alias IPs in the GKE cluster. If `true`, a VPC-native cluster is created. This field is only supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. Environments in newer versions always use VPC-native GKE clusters. }, - "location": "A String", # Optional. The Compute Engine [zone](/compute/docs/regions-zones) in which to deploy the VMs used to run the Apache Airflow software, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}". This `location` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.machineType` are specified, `nodeConfig.machineType` must belong to this `location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If only one field (`location` or `nodeConfig.machineType`) is specified, the location information from the specified field will be propagated to the unspecified field. - "machineType": "A String", # Optional. The Compute Engine [machine type](/compute/docs/machine-types) used for cluster instances, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}/machineTypes/{machineTypeId}". The `machineType` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.location` are specified, this `machineType` must belong to the `nodeConfig.location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If exactly one of this field and `nodeConfig.location` is specified, the location information from the specified field will be propagated to the unspecified field. The `machineTypeId` must not be a [shared-core machine type](/compute/docs/machine-types#sharedcore). If this field is unspecified, the `machineTypeId` defaults to "n1-standard-1". - "maxPodsPerNode": 42, # Optional. The maximum number of pods per node in the Cloud Composer GKE cluster. The value must be between 8 and 110 and it can be set only if the environment is VPC-native. The default value is 32. Values of this field will be propagated both to the `default-pool` node pool of the newly created GKE cluster, and to the default "Maximum Pods per Node" value which is used for newly created node pools if their value is not explicitly set during node pool creation. For more information, see [Optimizing IP address allocation] (https://cloud.google.com/kubernetes-engine/docs/how-to/flexible-pod-cidr). Cannot be updated. + "location": "A String", # Optional. The Compute Engine [zone](/compute/docs/regions-zones) in which to deploy the VMs used to run the Apache Airflow software, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}". This `location` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.machineType` are specified, `nodeConfig.machineType` must belong to this `location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If only one field (`location` or `nodeConfig.machineType`) is specified, the location information from the specified field will be propagated to the unspecified field. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "machineType": "A String", # Optional. The Compute Engine [machine type](/compute/docs/machine-types) used for cluster instances, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/zones/{zoneId}/machineTypes/{machineTypeId}". The `machineType` must belong to the enclosing environment's project and location. If both this field and `nodeConfig.location` are specified, this `machineType` must belong to the `nodeConfig.location`; if both are unspecified, the service will pick a zone in the Compute Engine region corresponding to the Cloud Composer location, and propagate that choice to both fields. If exactly one of this field and `nodeConfig.location` is specified, the location information from the specified field will be propagated to the unspecified field. The `machineTypeId` must not be a [shared-core machine type](/compute/docs/machine-types#sharedcore). If this field is unspecified, the `machineTypeId` defaults to "n1-standard-1". This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "maxPodsPerNode": 42, # Optional. The maximum number of pods per node in the Cloud Composer GKE cluster. The value must be between 8 and 110 and it can be set only if the environment is VPC-native. The default value is 32. Values of this field will be propagated both to the `default-pool` node pool of the newly created GKE cluster, and to the default "Maximum Pods per Node" value which is used for newly created node pools if their value is not explicitly set during node pool creation. For more information, see [Optimizing IP address allocation] (https://cloud.google.com/kubernetes-engine/docs/how-to/flexible-pod-cidr). Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "network": "A String", # Optional. The Compute Engine network to be used for machine communications, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/global/networks/{networkId}". If unspecified, the default network in the environment's project is used. If a [Custom Subnet Network](/vpc/docs/vpc#vpc_networks_and_subnets) is provided, `nodeConfig.subnetwork` must also be provided. For [Shared VPC](/vpc/docs/shared-vpc) subnetwork requirements, see `nodeConfig.subnetwork`. - "oauthScopes": [ # Optional. The set of Google API scopes to be made available on all node VMs. If `oauth_scopes` is empty, defaults to ["https://www.googleapis.com/auth/cloud-platform"]. Cannot be updated. + "oauthScopes": [ # Optional. The set of Google API scopes to be made available on all node VMs. If `oauth_scopes` is empty, defaults to ["https://www.googleapis.com/auth/cloud-platform"]. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "A String", ], "serviceAccount": "A String", # Optional. The Google Cloud Platform Service Account to be used by the workloads. If a service account is not specified, the "default" Compute Engine service account is used. Cannot be updated. "subnetwork": "A String", # Optional. The Compute Engine subnetwork to be used for machine communications, specified as a [relative resource name](/apis/design/resource_names#relative_resource_name). For example: "projects/{projectId}/regions/{regionId}/subnetworks/{subnetworkId}" If a subnetwork is provided, `nodeConfig.network` must also be provided, and the subnetwork must belong to the enclosing environment's project and location. - "tags": [ # Optional. The list of instance tags applied to all node VMs. Tags are used to identify valid sources or targets for network firewalls. Each tag within the list must comply with [RFC1035](https://www.ietf.org/rfc/rfc1035.txt). Cannot be updated. + "tags": [ # Optional. The list of instance tags applied to all node VMs. Tags are used to identify valid sources or targets for network firewalls. Each tag within the list must comply with [RFC1035](https://www.ietf.org/rfc/rfc1035.txt). Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "A String", ], }, - "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. + "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudSqlIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range in tenant project will be reserved for Cloud SQL. Needs to be disjoint from web_server_ipv4_cidr_block - "enablePrivateEnvironment": True or False, # Optional. If `true`, a Private IP Cloud Composer environment is created. If this field is set to true, `IPAllocationPolicy.use_ip_aliases` must be set to true . + "enablePrivateEnvironment": True or False, # Optional. If `true`, a Private IP Cloud Composer environment is created. If this field is set to true, `IPAllocationPolicy.use_ip_aliases` must be set to true for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "enablePrivatelyUsedPublicIps": True or False, # Optional. When enabled, IPs from public (non-RFC1918) ranges can be used for `IPAllocationPolicy.cluster_ipv4_cidr_block` and `IPAllocationPolicy.service_ipv4_cidr_block`. "privateClusterConfig": { # Configuration options for the private GKE cluster in a Cloud Composer environment. # Optional. Configuration for the private GKE cluster for a Private IP Cloud Composer environment. "enablePrivateEndpoint": True or False, # Optional. If `true`, access to the public endpoint of the GKE cluster is denied. "masterIpv4CidrBlock": "A String", # Optional. The CIDR block from which IPv4 range for GKE master will be reserved. If left blank, the default value of '172.16.0.0/23' is used. "masterIpv4ReservedRange": "A String", # Output only. The IP range in CIDR notation to use for the hosted master network. This range is used for assigning internal IP addresses to the cluster master or set of masters and to the internal load balancer virtual IP. This range must not overlap with any other ranges in use within the cluster's network. }, - "webServerIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for web server will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. - "webServerIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's App Engine VMs. + "webServerIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for web server will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's App Engine VMs. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. }, "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. @@ -673,13 +681,13 @@Method Details
"pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, - "pythonVersion": "A String", # Optional. The major version of Python used to run the Apache Airflow scheduler, worker, and webserver processes. Can be set to '2' or '3'. If not specified, the default is '3'. Cannot be updated. - "schedulerCount": 42, # Optional. The number of schedulers for Airflow. + "pythonVersion": "A String", # Optional. The major version of Python used to run the Apache Airflow scheduler, worker, and webserver processes. Can be set to '2' or '3'. If not specified, the default is '3'. Cannot be updated. This field is only supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. Environments in newer versions always use Python major version 3. + "schedulerCount": 42, # Optional. The number of schedulers for Airflow. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-2.*.*. }, - "webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. # Optional. The configuration settings for the Airflow web server App Engine instance. + "webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for the Airflow web server App Engine instance. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. @@ -718,7 +726,7 @@Method Details
"uuid": "A String", # Output only. The UUID (Universally Unique IDentifier) associated with this environment. This value is generated when the environment is created. } - updateMask: string, Required. A comma-separated list of paths, relative to `Environment`, of fields to update. For example, to set the version of scikit-learn to install in the environment to 0.19.0 and to remove an existing installation of argparse, the `updateMask` parameter would include the following two `paths` values: "config.softwareConfig.pypiPackages.scikit-learn" and "config.softwareConfig.pypiPackages.argparse". The included patch environment would specify the scikit-learn version as follows: { "config":{ "softwareConfig":{ "pypiPackages":{ "scikit-learn":"==0.19.0" } } } } Note that in the above example, any existing PyPI packages other than scikit-learn and argparse will be unaffected. Only one update type may be included in a single request's `updateMask`. For example, one cannot update both the PyPI packages and labels in the same request. However, it is possible to update multiple members of a map field simultaneously in the same request. For example, to set the labels "label1" and "label2" while clearing "label3" (assuming it already exists), one can provide the paths "labels.label1", "labels.label2", and "labels.label3" and populate the patch environment as follows: { "labels":{ "label1":"new-label1-value" "label2":"new-label2-value" } } Note that in the above example, any existing labels that are not included in the `updateMask` will be unaffected. It is also possible to replace an entire map field by providing the map field's path in the `updateMask`. The new value of the field will be that which is provided in the patch environment. For example, to delete all pre-existing user-specified PyPI packages and install botocore at version 1.7.14, the `updateMask` would contain the path "config.softwareConfig.pypiPackages", and the patch environment would be the following: { "config":{ "softwareConfig":{ "pypiPackages":{ "botocore":"==1.7.14" } } } } **Note:** Only the following fields can be updated: * `config.softwareConfig.pypiPackages` * Replace all custom custom PyPI packages. If a replacement package map is not included in `environment`, all custom PyPI packages are cleared. It is an error to provide both this mask and a mask specifying an individual package. * `config.softwareConfig.pypiPackages.`packagename * Update the custom PyPI package *packagename*, preserving other packages. To delete the package, include it in `updateMask`, and omit the mapping for it in `environment.config.softwareConfig.pypiPackages`. It is an error to provide both a mask of this form and the `config.softwareConfig.pypiPackages` mask. * `labels` * Replace all environment labels. If a replacement labels map is not included in `environment`, all labels are cleared. It is an error to provide both this mask and a mask specifying one or more individual labels. * `labels.`labelName * Set the label named *labelName*, while preserving other labels. To delete the label, include it in `updateMask` and omit its mapping in `environment.labels`. It is an error to provide both a mask of this form and the `labels` mask. * `config.nodeCount` * Horizontally scale the number of nodes in the environment. An integer greater than or equal to 3 must be provided in the `config.nodeCount` field. * `config.webServerNetworkAccessControl` * Replace the environment's current WebServerNetworkAccessControl. * `config.softwareConfig.airflowConfigOverrides` * Replace all Apache Airflow config overrides. If a replacement config overrides map is not included in `environment`, all config overrides are cleared. It is an error to provide both this mask and a mask specifying one or more individual config overrides. * `config.softwareConfig.airflowConfigOverrides.`section-name * Override the Apache Airflow config property *name* in the section named *section*, preserving other properties. To delete the property override, include it in `updateMask` and omit its mapping in `environment.config.softwareConfig.airflowConfigOverrides`. It is an error to provide both a mask of this form and the `config.softwareConfig.airflowConfigOverrides` mask. * `config.softwareConfig.envVariables` * Replace all environment variables. If a replacement environment variable map is not included in `environment`, all custom environment variables are cleared. It is an error to provide both this mask and a mask specifying one or more individual environment variables. * `config.softwareConfig.imageVersion` * Upgrade the version of the environment in-place. Refer to `SoftwareConfig.image_version` for information on how to format the new image version. Additionally, the new image version cannot effect a version downgrade and must match the current image version's Composer major version and Airflow major and minor versions. Consult the [Cloud Composer Version List](https://cloud.google.com/composer/docs/concepts/versioning/composer-versions) for valid values. * `config.softwareConfig.schedulerCount` * Horizontally scale the number of schedulers in Airflow. A positive integer not greater than the number of nodes must be provided in the `config.softwareConfig.schedulerCount` field. * `config.databaseConfig.machineType` * Cloud SQL machine type used by Airflow database. It has to be one of: db-n1-standard-2, db-n1-standard-4, db-n1-standard-8 or db-n1-standard-16. * `config.webServerConfig.machineType` * Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. * `config.maintenanceWindow` * Maintenance window during which Cloud Composer components may be under maintenance. + updateMask: string, Required. A comma-separated list of paths, relative to `Environment`, of fields to update. For example, to set the version of scikit-learn to install in the environment to 0.19.0 and to remove an existing installation of argparse, the `updateMask` parameter would include the following two `paths` values: "config.softwareConfig.pypiPackages.scikit-learn" and "config.softwareConfig.pypiPackages.argparse". The included patch environment would specify the scikit-learn version as follows: { "config":{ "softwareConfig":{ "pypiPackages":{ "scikit-learn":"==0.19.0" } } } } Note that in the above example, any existing PyPI packages other than scikit-learn and argparse will be unaffected. Only one update type may be included in a single request's `updateMask`. For example, one cannot update both the PyPI packages and labels in the same request. However, it is possible to update multiple members of a map field simultaneously in the same request. For example, to set the labels "label1" and "label2" while clearing "label3" (assuming it already exists), one can provide the paths "labels.label1", "labels.label2", and "labels.label3" and populate the patch environment as follows: { "labels":{ "label1":"new-label1-value" "label2":"new-label2-value" } } Note that in the above example, any existing labels that are not included in the `updateMask` will be unaffected. It is also possible to replace an entire map field by providing the map field's path in the `updateMask`. The new value of the field will be that which is provided in the patch environment. For example, to delete all pre-existing user-specified PyPI packages and install botocore at version 1.7.14, the `updateMask` would contain the path "config.softwareConfig.pypiPackages", and the patch environment would be the following: { "config":{ "softwareConfig":{ "pypiPackages":{ "botocore":"==1.7.14" } } } } **Note:** Only the following fields can be updated: * `config.softwareConfig.pypiPackages` * Replace all custom custom PyPI packages. If a replacement package map is not included in `environment`, all custom PyPI packages are cleared. It is an error to provide both this mask and a mask specifying an individual package. * `config.softwareConfig.pypiPackages.`packagename * Update the custom PyPI package *packagename*, preserving other packages. To delete the package, include it in `updateMask`, and omit the mapping for it in `environment.config.softwareConfig.pypiPackages`. It is an error to provide both a mask of this form and the `config.softwareConfig.pypiPackages` mask. * `labels` * Replace all environment labels. If a replacement labels map is not included in `environment`, all labels are cleared. It is an error to provide both this mask and a mask specifying one or more individual labels. * `labels.`labelName * Set the label named *labelName*, while preserving other labels. To delete the label, include it in `updateMask` and omit its mapping in `environment.labels`. It is an error to provide both a mask of this form and the `labels` mask. * `config.nodeCount` * Horizontally scale the number of nodes in the environment. An integer greater than or equal to 3 must be provided in the `config.nodeCount` field. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. * `config.webServerNetworkAccessControl` * Replace the environment's current WebServerNetworkAccessControl. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. * `config.softwareConfig.airflowConfigOverrides` * Replace all Apache Airflow config overrides. If a replacement config overrides map is not included in `environment`, all config overrides are cleared. It is an error to provide both this mask and a mask specifying one or more individual config overrides. * `config.softwareConfig.airflowConfigOverrides.`section-name * Override the Apache Airflow config property *name* in the section named *section*, preserving other properties. To delete the property override, include it in `updateMask` and omit its mapping in `environment.config.softwareConfig.airflowConfigOverrides`. It is an error to provide both a mask of this form and the `config.softwareConfig.airflowConfigOverrides` mask. * `config.softwareConfig.envVariables` * Replace all environment variables. If a replacement environment variable map is not included in `environment`, all custom environment variables are cleared. It is an error to provide both this mask and a mask specifying one or more individual environment variables. * `config.softwareConfig.imageVersion` * Upgrade the version of the environment in-place. Refer to `SoftwareConfig.image_version` for information on how to format the new image version. Additionally, the new image version cannot effect a version downgrade and must match the current image version's Composer major version and Airflow major and minor versions. Consult the [Cloud Composer Version List](https://cloud.google.com/composer/docs/concepts/versioning/composer-versions) for valid values. * `config.softwareConfig.schedulerCount` * Horizontally scale the number of schedulers in Airflow. A positive integer not greater than the number of nodes must be provided in the `config.softwareConfig.schedulerCount` field. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-2.*.*. * `config.databaseConfig.machineType` * Cloud SQL machine type used by Airflow database. It has to be one of: db-n1-standard-2, db-n1-standard-4, db-n1-standard-8 or db-n1-standard-16. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. * `config.webServerConfig.machineType` * Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. * `config.maintenanceWindow` * Maintenance window during which Cloud Composer components may be under maintenance. * `config.workloadsConfig` * The workloads configuration settings for the GKE cluster associated with the Cloud Composer environment. Supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. * `config.environmentSize` * The size of the Cloud Composer environment. Supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. x__xgafv: string, V1 error format. Allowed values 1 - v1 error format diff --git a/docs/dyn/contactcenterinsights_v1.html b/docs/dyn/contactcenterinsights_v1.html new file mode 100644 index 00000000000..d405fd7d8dc --- /dev/null +++ b/docs/dyn/contactcenterinsights_v1.html @@ -0,0 +1,111 @@ + + + +Contact Center AI Insights API
+Instance Methods
++
+projects()+Returns the projects Resource.
+ ++
+close()Close httplib2 connections.
+ +Create a BatchHttpRequest object based on the discovery document.
+Method Details
+++ +close()+Close httplib2 connections.+++ + \ No newline at end of file diff --git a/docs/dyn/contactcenterinsights_v1.projects.html b/docs/dyn/contactcenterinsights_v1.projects.html new file mode 100644 index 00000000000..517e91b9d29 --- /dev/null +++ b/docs/dyn/contactcenterinsights_v1.projects.html @@ -0,0 +1,91 @@ + + + +new_batch_http_request()+Create a BatchHttpRequest object based on the discovery document. + + Args: + callback: callable, A callback to be called for each response, of the + form callback(id, response, exception). The first parameter is the + request id, and the second is the deserialized response object. The + third is an apiclient.errors.HttpError exception object if an HTTP + error occurred while processing the request, or None if no error + occurred. + + Returns: + A BatchHttpRequest object based on the discovery document. ++Contact Center AI Insights API . projects
+Instance Methods
++
+locations()+Returns the locations Resource.
+ ++
+close()Close httplib2 connections.
+Method Details
+++ + \ No newline at end of file diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.analyses.html b/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.analyses.html new file mode 100644 index 00000000000..fde901124d6 --- /dev/null +++ b/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.analyses.html @@ -0,0 +1,488 @@ + + + +close()+Close httplib2 connections.+Contact Center AI Insights API . projects . locations . conversations . analyses
+Instance Methods
++
+close()Close httplib2 connections.
++
+create(parent, body=None, x__xgafv=None)Creates an analysis. The long running operation is done when the analysis has completed.
+ +Deletes an analysis.
+ +Gets an analysis.
++
+list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)Lists analyses.
++
+list_next(previous_request, previous_response)Retrieves the next page of results.
+Method Details
+++ +close()+Close httplib2 connections.+++ +create(parent, body=None, x__xgafv=None)+Creates an analysis. The long running operation is done when the analysis has completed. + +Args: + parent: string, Required. The parent resource of the analysis. (required) + body: object, The request body. + The object takes the form of: + +{ # The analysis resource. + "analysisResult": { # The result of an analysis. # Output only. The result of the analysis, which is populated when the analysis finishes. + "callAnalysisMetadata": { # Call-specific metadata created during analysis. # Call-specific metadata created by the analysis. + "annotations": [ # A list of call annotations that apply to this call. + { # A piece of metadata that applies to a window of a call. + "annotationEndBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "annotationStartBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "channelTag": 42, # The channel of the audio where the annotation occurs. For single-channel audio, this field is not populated. + "entityMentionData": { # The data for an entity mention annotation. This represents a mention of an `Entity` in the conversation. # Data specifying an entity mention. + "entityUniqueId": "A String", # The key of this entity in conversation entities. Can be used to retrieve the exact `Entity` this mention is attached to. + "sentiment": { # The data for a sentiment annotation. # Sentiment expressed for this mention of the entity. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The type of the entity mention. + }, + "holdData": { # The data for a hold annotation. # Data specifying a hold. + }, + "intentMatchData": { # The data for an intent match. Represents an intent match for a text segment in the conversation. A text segment can be part of a sentence, a complete sentence, or an utterance with multiple sentences. # Data specifying an intent match. + "intentUniqueId": "A String", # The id of the matched intent. Can be used to retrieve the corresponding intent information. + }, + "interruptionData": { # The data for an interruption annotation. # Data specifying an interruption. + }, + "phraseMatchData": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. # Data specifying a phrase match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "silenceData": { # The data for a silence annotation. # Data specifying silence. + }, + }, + ], + "entities": { # All the entities in the call. + "a_key": { # The data for an entity annotation. Represents a phrase in the conversation that is a known entity, such as a person, an organization, or location. + "displayName": "A String", # The representative name for the entity. + "metadata": { # Metadata associated with the entity. For most entity types, the metadata is a Wikipedia URL (`wikipedia_url`) and Knowledge Graph MID (`mid`), if they are available. For the metadata associated with other entity types, see the Type table below. + "a_key": "A String", + }, + "salience": 3.14, # The salience score associated with the entity in the [0, 1.0] range. The salience score for an entity provides information about the importance or centrality of that entity to the entire document text. Scores closer to 0 are less salient, while scores closer to 1.0 are highly salient. + "sentiment": { # The data for a sentiment annotation. # The aggregate sentiment expressed for this entity in the conversation. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The entity type. + }, + }, + "intents": { # All the matched intents in the call. + "a_key": { # The data for an intent. Represents a detected intent in the conversation, for example MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + "id": "A String", # The unique identifier of the intent. + }, + }, + "issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result. + "issueModel": "A String", # Issue model that generates the result. + "issues": [ # All the matched issues. + { # Information about the issue. + "issue": "A String", # Resource name of the assigned issue. + "score": 3.14, # Score indicating the likelihood of the issue assignment. currently bounded on [0,1]. + }, + ], + }, + "phraseMatchers": { # All the matched phrase matchers in the call. + "a_key": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + }, + "sentiments": [ # Overall conversation-level sentiment for each channel of the call. + { # One channel of conversation-level sentiment data. + "channelTag": 42, # The channel of the audio that the data applies to. + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + }, + ], + }, + "endTime": "A String", # The time at which the analysis ended. + }, + "createTime": "A String", # Output only. The time at which the analysis was created, which occurs when the long-running operation completes. + "name": "A String", # Immutable. The resource name of the analysis. Format: projects/{project}/locations/{location}/conversations/{conversation}/analyses/{analysis} + "requestTime": "A String", # Output only. The time at which the analysis was requested. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +delete(name, x__xgafv=None)+Deletes an analysis. + +Args: + name: string, Required. The name of the analysis to delete. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`. +}+++ +get(name, x__xgafv=None)+Gets an analysis. + +Args: + name: string, Required. The name of the analysis to get. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The analysis resource. + "analysisResult": { # The result of an analysis. # Output only. The result of the analysis, which is populated when the analysis finishes. + "callAnalysisMetadata": { # Call-specific metadata created during analysis. # Call-specific metadata created by the analysis. + "annotations": [ # A list of call annotations that apply to this call. + { # A piece of metadata that applies to a window of a call. + "annotationEndBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "annotationStartBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "channelTag": 42, # The channel of the audio where the annotation occurs. For single-channel audio, this field is not populated. + "entityMentionData": { # The data for an entity mention annotation. This represents a mention of an `Entity` in the conversation. # Data specifying an entity mention. + "entityUniqueId": "A String", # The key of this entity in conversation entities. Can be used to retrieve the exact `Entity` this mention is attached to. + "sentiment": { # The data for a sentiment annotation. # Sentiment expressed for this mention of the entity. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The type of the entity mention. + }, + "holdData": { # The data for a hold annotation. # Data specifying a hold. + }, + "intentMatchData": { # The data for an intent match. Represents an intent match for a text segment in the conversation. A text segment can be part of a sentence, a complete sentence, or an utterance with multiple sentences. # Data specifying an intent match. + "intentUniqueId": "A String", # The id of the matched intent. Can be used to retrieve the corresponding intent information. + }, + "interruptionData": { # The data for an interruption annotation. # Data specifying an interruption. + }, + "phraseMatchData": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. # Data specifying a phrase match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "silenceData": { # The data for a silence annotation. # Data specifying silence. + }, + }, + ], + "entities": { # All the entities in the call. + "a_key": { # The data for an entity annotation. Represents a phrase in the conversation that is a known entity, such as a person, an organization, or location. + "displayName": "A String", # The representative name for the entity. + "metadata": { # Metadata associated with the entity. For most entity types, the metadata is a Wikipedia URL (`wikipedia_url`) and Knowledge Graph MID (`mid`), if they are available. For the metadata associated with other entity types, see the Type table below. + "a_key": "A String", + }, + "salience": 3.14, # The salience score associated with the entity in the [0, 1.0] range. The salience score for an entity provides information about the importance or centrality of that entity to the entire document text. Scores closer to 0 are less salient, while scores closer to 1.0 are highly salient. + "sentiment": { # The data for a sentiment annotation. # The aggregate sentiment expressed for this entity in the conversation. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The entity type. + }, + }, + "intents": { # All the matched intents in the call. + "a_key": { # The data for an intent. Represents a detected intent in the conversation, for example MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + "id": "A String", # The unique identifier of the intent. + }, + }, + "issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result. + "issueModel": "A String", # Issue model that generates the result. + "issues": [ # All the matched issues. + { # Information about the issue. + "issue": "A String", # Resource name of the assigned issue. + "score": 3.14, # Score indicating the likelihood of the issue assignment. currently bounded on [0,1]. + }, + ], + }, + "phraseMatchers": { # All the matched phrase matchers in the call. + "a_key": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + }, + "sentiments": [ # Overall conversation-level sentiment for each channel of the call. + { # One channel of conversation-level sentiment data. + "channelTag": 42, # The channel of the audio that the data applies to. + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + }, + ], + }, + "endTime": "A String", # The time at which the analysis ended. + }, + "createTime": "A String", # Output only. The time at which the analysis was created, which occurs when the long-running operation completes. + "name": "A String", # Immutable. The resource name of the analysis. Format: projects/{project}/locations/{location}/conversations/{conversation}/analyses/{analysis} + "requestTime": "A String", # Output only. The time at which the analysis was requested. +}+++ +list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)+Lists analyses. + +Args: + parent: string, Required. The parent resource of the analyses. (required) + filter: string, A filter to reduce results to a specific subset. Useful for querying conversations with specific properties. + pageSize: integer, The maximum number of analyses to return in the response. If this value is zero, the service will select a default size. A call might return fewer objects than requested. A non-empty `next_page_token` in the response indicates that more data is available. + pageToken: string, The value returned by the last `ListAnalysesResponse`; indicates that this is a continuation of a prior `ListAnalyses` call and the system should return the next page of data. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response to list analyses. + "analyses": [ # The analyses that match the request. + { # The analysis resource. + "analysisResult": { # The result of an analysis. # Output only. The result of the analysis, which is populated when the analysis finishes. + "callAnalysisMetadata": { # Call-specific metadata created during analysis. # Call-specific metadata created by the analysis. + "annotations": [ # A list of call annotations that apply to this call. + { # A piece of metadata that applies to a window of a call. + "annotationEndBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "annotationStartBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "channelTag": 42, # The channel of the audio where the annotation occurs. For single-channel audio, this field is not populated. + "entityMentionData": { # The data for an entity mention annotation. This represents a mention of an `Entity` in the conversation. # Data specifying an entity mention. + "entityUniqueId": "A String", # The key of this entity in conversation entities. Can be used to retrieve the exact `Entity` this mention is attached to. + "sentiment": { # The data for a sentiment annotation. # Sentiment expressed for this mention of the entity. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The type of the entity mention. + }, + "holdData": { # The data for a hold annotation. # Data specifying a hold. + }, + "intentMatchData": { # The data for an intent match. Represents an intent match for a text segment in the conversation. A text segment can be part of a sentence, a complete sentence, or an utterance with multiple sentences. # Data specifying an intent match. + "intentUniqueId": "A String", # The id of the matched intent. Can be used to retrieve the corresponding intent information. + }, + "interruptionData": { # The data for an interruption annotation. # Data specifying an interruption. + }, + "phraseMatchData": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. # Data specifying a phrase match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "silenceData": { # The data for a silence annotation. # Data specifying silence. + }, + }, + ], + "entities": { # All the entities in the call. + "a_key": { # The data for an entity annotation. Represents a phrase in the conversation that is a known entity, such as a person, an organization, or location. + "displayName": "A String", # The representative name for the entity. + "metadata": { # Metadata associated with the entity. For most entity types, the metadata is a Wikipedia URL (`wikipedia_url`) and Knowledge Graph MID (`mid`), if they are available. For the metadata associated with other entity types, see the Type table below. + "a_key": "A String", + }, + "salience": 3.14, # The salience score associated with the entity in the [0, 1.0] range. The salience score for an entity provides information about the importance or centrality of that entity to the entire document text. Scores closer to 0 are less salient, while scores closer to 1.0 are highly salient. + "sentiment": { # The data for a sentiment annotation. # The aggregate sentiment expressed for this entity in the conversation. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The entity type. + }, + }, + "intents": { # All the matched intents in the call. + "a_key": { # The data for an intent. Represents a detected intent in the conversation, for example MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + "id": "A String", # The unique identifier of the intent. + }, + }, + "issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result. + "issueModel": "A String", # Issue model that generates the result. + "issues": [ # All the matched issues. + { # Information about the issue. + "issue": "A String", # Resource name of the assigned issue. + "score": 3.14, # Score indicating the likelihood of the issue assignment. currently bounded on [0,1]. + }, + ], + }, + "phraseMatchers": { # All the matched phrase matchers in the call. + "a_key": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + }, + "sentiments": [ # Overall conversation-level sentiment for each channel of the call. + { # One channel of conversation-level sentiment data. + "channelTag": 42, # The channel of the audio that the data applies to. + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + }, + ], + }, + "endTime": "A String", # The time at which the analysis ended. + }, + "createTime": "A String", # Output only. The time at which the analysis was created, which occurs when the long-running operation completes. + "name": "A String", # Immutable. The resource name of the analysis. Format: projects/{project}/locations/{location}/conversations/{conversation}/analyses/{analysis} + "requestTime": "A String", # Output only. The time at which the analysis was requested. + }, + ], + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. +}+++ + \ No newline at end of file diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.html b/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.html new file mode 100644 index 00000000000..bf74afb330d --- /dev/null +++ b/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.html @@ -0,0 +1,1525 @@ + + + +list_next(previous_request, previous_response)+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Contact Center AI Insights API . projects . locations . conversations
+Instance Methods
++
+analyses()+Returns the analyses Resource.
+ ++
+calculateStats(location, filter=None, x__xgafv=None)Gets conversation statistics.
++
+close()Close httplib2 connections.
++
+create(parent, body=None, conversationId=None, x__xgafv=None)Creates a conversation.
++
+delete(name, force=None, x__xgafv=None)Deletes a conversation.
++
+get(name, view=None, x__xgafv=None)Gets a conversation.
++
+list(parent, filter=None, pageSize=None, pageToken=None, view=None, x__xgafv=None)Lists conversations.
++
+list_next(previous_request, previous_response)Retrieves the next page of results.
++
+patch(name, body=None, updateMask=None, x__xgafv=None)Updates a conversation.
+Method Details
+++ +calculateStats(location, filter=None, x__xgafv=None)+Gets conversation statistics. + +Args: + location: string, Required. The location of the conversations. (required) + filter: string, A filter to reduce results to a specific subset. This field is useful for getting statistics about conversations with specific properties. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response for calculating conversation statistics. + "averageDuration": "A String", # The average duration of all conversations. The average is calculated using only conversations that have a time duration. + "averageTurnCount": 42, # The average number of turns per conversation. + "conversationCount": 42, # The total number of conversations. + "conversationCountTimeSeries": { # A time series representing conversations over time. # A time series representing the count of conversations created over time that match that requested filter criteria. + "intervalDuration": "A String", # The duration of each interval. + "points": [ # An ordered list of intervals from earliest to latest, where each interval represents the number of conversations that transpired during the time window. + { # A single interval in a time series. + "conversationCount": 42, # The number of conversations created in this interval. + "startTime": "A String", # The start time of this interval. + }, + ], + }, + "customHighlighterMatches": { # A map associating each custom highlighter resource name with its respective number of matches in the set of conversations. + "a_key": 42, + }, + "issueMatches": { # A map associating each issue resource name with its respective number of matches in the set of conversations. Key has the format: `projects//locations//issueModels//issues/` + "a_key": 42, + }, + "smartHighlighterMatches": { # A map associating each smart highlighter display name with its respective number of matches in the set of conversations. + "a_key": 42, + }, +}+++ +close()+Close httplib2 connections.+++ +create(parent, body=None, conversationId=None, x__xgafv=None)+Creates a conversation. + +Args: + parent: string, Required. The parent resource of the conversation. (required) + body: object, The request body. + The object takes the form of: + +{ # The conversation resource. + "agentId": "A String", # An opaque, user-specified string representing the human agent who handled the conversation. + "callMetadata": { # Call-specific metadata. # Call-specific metadata. + "agentChannel": 42, # The audio channel that contains the agent. + "customerChannel": 42, # The audio channel that contains the customer. + }, + "createTime": "A String", # Output only. The time at which the conversation was created. + "dataSource": { # The conversation source, which is a combination of transcript and audio. # The source of the audio and transcription for the conversation. + "dialogflowSource": { # A Dialogflow source of conversation data. # The source when the conversation comes from Dialogflow. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "dialogflowConversation": "A String", # Output only. The name of the Dialogflow conversation that this conversation resource is derived from. Format: projects/{project}/locations/{location}/conversations/{conversation} + }, + "gcsSource": { # A Cloud Storage source of conversation data. # A Cloud Storage location specification for the audio and transcript. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "transcriptUri": "A String", # Immutable. Cloud Storage URI that points to a file that contains the conversation transcript. + }, + }, + "dialogflowIntents": { # Output only. All the matched Dialogflow intents in the call. The key corresponds to a Dialogflow intent, format: projects/{project}/agent/{agent}/intents/{intent} + "a_key": { # The data for a Dialogflow intent. Represents a detected intent in the conversation, e.g. MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + }, + }, + "duration": "A String", # Output only. The duration of the conversation. + "expireTime": "A String", # The time at which this conversation should expire. After this time, the conversation data and any associated analyses will be deleted. + "labels": { # A map for the user to specify any custom fields. A maximum of 20 labels per conversation is allowed, with a maximum of 256 characters per entry. + "a_key": "A String", + }, + "languageCode": "A String", # A user-specified language code for the conversation. + "latestAnalysis": { # The analysis resource. # Output only. The conversation's latest analysis, if one exists. + "analysisResult": { # The result of an analysis. # Output only. The result of the analysis, which is populated when the analysis finishes. + "callAnalysisMetadata": { # Call-specific metadata created during analysis. # Call-specific metadata created by the analysis. + "annotations": [ # A list of call annotations that apply to this call. + { # A piece of metadata that applies to a window of a call. + "annotationEndBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "annotationStartBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "channelTag": 42, # The channel of the audio where the annotation occurs. For single-channel audio, this field is not populated. + "entityMentionData": { # The data for an entity mention annotation. This represents a mention of an `Entity` in the conversation. # Data specifying an entity mention. + "entityUniqueId": "A String", # The key of this entity in conversation entities. Can be used to retrieve the exact `Entity` this mention is attached to. + "sentiment": { # The data for a sentiment annotation. # Sentiment expressed for this mention of the entity. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The type of the entity mention. + }, + "holdData": { # The data for a hold annotation. # Data specifying a hold. + }, + "intentMatchData": { # The data for an intent match. Represents an intent match for a text segment in the conversation. A text segment can be part of a sentence, a complete sentence, or an utterance with multiple sentences. # Data specifying an intent match. + "intentUniqueId": "A String", # The id of the matched intent. Can be used to retrieve the corresponding intent information. + }, + "interruptionData": { # The data for an interruption annotation. # Data specifying an interruption. + }, + "phraseMatchData": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. # Data specifying a phrase match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "silenceData": { # The data for a silence annotation. # Data specifying silence. + }, + }, + ], + "entities": { # All the entities in the call. + "a_key": { # The data for an entity annotation. Represents a phrase in the conversation that is a known entity, such as a person, an organization, or location. + "displayName": "A String", # The representative name for the entity. + "metadata": { # Metadata associated with the entity. For most entity types, the metadata is a Wikipedia URL (`wikipedia_url`) and Knowledge Graph MID (`mid`), if they are available. For the metadata associated with other entity types, see the Type table below. + "a_key": "A String", + }, + "salience": 3.14, # The salience score associated with the entity in the [0, 1.0] range. The salience score for an entity provides information about the importance or centrality of that entity to the entire document text. Scores closer to 0 are less salient, while scores closer to 1.0 are highly salient. + "sentiment": { # The data for a sentiment annotation. # The aggregate sentiment expressed for this entity in the conversation. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The entity type. + }, + }, + "intents": { # All the matched intents in the call. + "a_key": { # The data for an intent. Represents a detected intent in the conversation, for example MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + "id": "A String", # The unique identifier of the intent. + }, + }, + "issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result. + "issueModel": "A String", # Issue model that generates the result. + "issues": [ # All the matched issues. + { # Information about the issue. + "issue": "A String", # Resource name of the assigned issue. + "score": 3.14, # Score indicating the likelihood of the issue assignment. currently bounded on [0,1]. + }, + ], + }, + "phraseMatchers": { # All the matched phrase matchers in the call. + "a_key": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + }, + "sentiments": [ # Overall conversation-level sentiment for each channel of the call. + { # One channel of conversation-level sentiment data. + "channelTag": 42, # The channel of the audio that the data applies to. + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + }, + ], + }, + "endTime": "A String", # The time at which the analysis ended. + }, + "createTime": "A String", # Output only. The time at which the analysis was created, which occurs when the long-running operation completes. + "name": "A String", # Immutable. The resource name of the analysis. Format: projects/{project}/locations/{location}/conversations/{conversation}/analyses/{analysis} + "requestTime": "A String", # Output only. The time at which the analysis was requested. + }, + "medium": "A String", # Immutable. The conversation medium. + "name": "A String", # Immutable. The resource name of the conversation. Format: projects/{project}/locations/{location}/conversations/{conversation} + "runtimeAnnotations": [ # Output only. The annotations that were generated during the customer and agent interaction. + { # An annotation that was generated during the customer and agent interaction. + "annotationId": "A String", # The unique identifier of the annotation. Format: projects/{project}/locations/{location}/conversationDatasets/{dataset}/conversationDataItems/{data_item}/conversationAnnotations/{annotation} + "answerFeedback": { # The feedback that the customer has about a certain answer in the conversation. # The feedback that the customer has about the answer in `data`. + "clicked": True or False, # Indicates whether an answer or item was clicked by the human agent. + "correctnessLevel": "A String", # The correctness level of an answer. + "displayed": True or False, # Indicates whether an answer or item was displayed to the human agent in the agent desktop UI. + }, + "articleSuggestion": { # Agent Assist Article Suggestion data. # Agent Assist Article Suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this article is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Article Suggestion and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document} + "title": "A String", # Article title. + "uri": "A String", # Article URI. + }, + "createTime": "A String", # The time at which this annotation was created. + "dialogflowInteraction": { # Dialogflow interaction data. # Dialogflow interaction data. + "confidence": 3.14, # The confidence of the match ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "dialogflowIntentId": "A String", # The Dialogflow intent resource path. Format: projects/{project}/agent/{agent}/intents/{intent} + }, + "endBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "faqAnswer": { # Agent Assist frequently-asked-question answer data. # Agent Assist FAQ answer data. + "answer": "A String", # The piece of text from the `source` knowledge base document. + "confidenceScore": 3.14, # The system's confidence score that this answer is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the FAQ answer and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record}. + "question": "A String", # The corresponding FAQ question. + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. + }, + "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "suggestion": "A String", # The content of the suggestion. + }, + "smartReply": { # Agent Assist Smart Reply data. # Agent Assist Smart Reply data. + "confidenceScore": 3.14, # The system's confidence score that this reply is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Reply and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "reply": "A String", # The content of the reply. + }, + "startBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + }, + ], + "startTime": "A String", # The time at which the conversation started. + "transcript": { # A message representing the transcript of a conversation. # Output only. The conversation transcript. + "transcriptSegments": [ # A list of sequential transcript segments that comprise the conversation. + { # A segment of a full transcript. + "channelTag": 42, # For conversations derived from multi-channel audio, this is the channel number corresponding to the audio from that channel. For audioChannelCount = N, its output values can range from '1' to 'N'. A channel tag of 0 indicates that the audio is mono. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this segment. A default value of 0.0 indicates that the value is unset. + "languageCode": "A String", # The language code of this segment as a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US". + "segmentParticipant": { # The call participant speaking for a given utterance. # The participant of this segment. + "dialogflowParticipant": "A String", # The name of the Dialogflow participant. Format: projects/{project}/locations/{location}/conversations/{conversation}/participants/{participant} + "role": "A String", # The role of the participant. + }, + "text": "A String", # The text of this segment. + "words": [ # A list of the word-specific information for each word in the segment. + { # Word-level info for words in a transcript. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this word. A default value of 0.0 indicates that the value is unset. + "endOffset": "A String", # Time offset of the end of this word relative to the beginning of the total conversation. + "startOffset": "A String", # Time offset of the start of this word relative to the beginning of the total conversation. + "word": "A String", # The word itself. Includes punctuation marks that surround the word. + }, + ], + }, + ], + }, + "ttl": "A String", # Input only. The TTL for this resource. If specified, then this TTL will be used to calculate the expire time. + "turnCount": 42, # Output only. The number of turns in the conversation. + "updateTime": "A String", # Output only. The most recent time at which the conversation was updated. +} + + conversationId: string, A unique ID for the new conversation. This ID will become the final component of the conversation's resource name. If no ID is specified, a server-generated ID will be used. This value should be 4-32 characters and must match the regular expression /^[a-z0-9-]{4,32}$/. Valid characters are /a-z-/ + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The conversation resource. + "agentId": "A String", # An opaque, user-specified string representing the human agent who handled the conversation. + "callMetadata": { # Call-specific metadata. # Call-specific metadata. + "agentChannel": 42, # The audio channel that contains the agent. + "customerChannel": 42, # The audio channel that contains the customer. + }, + "createTime": "A String", # Output only. The time at which the conversation was created. + "dataSource": { # The conversation source, which is a combination of transcript and audio. # The source of the audio and transcription for the conversation. + "dialogflowSource": { # A Dialogflow source of conversation data. # The source when the conversation comes from Dialogflow. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "dialogflowConversation": "A String", # Output only. The name of the Dialogflow conversation that this conversation resource is derived from. Format: projects/{project}/locations/{location}/conversations/{conversation} + }, + "gcsSource": { # A Cloud Storage source of conversation data. # A Cloud Storage location specification for the audio and transcript. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "transcriptUri": "A String", # Immutable. Cloud Storage URI that points to a file that contains the conversation transcript. + }, + }, + "dialogflowIntents": { # Output only. All the matched Dialogflow intents in the call. The key corresponds to a Dialogflow intent, format: projects/{project}/agent/{agent}/intents/{intent} + "a_key": { # The data for a Dialogflow intent. Represents a detected intent in the conversation, e.g. MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + }, + }, + "duration": "A String", # Output only. The duration of the conversation. + "expireTime": "A String", # The time at which this conversation should expire. After this time, the conversation data and any associated analyses will be deleted. + "labels": { # A map for the user to specify any custom fields. A maximum of 20 labels per conversation is allowed, with a maximum of 256 characters per entry. + "a_key": "A String", + }, + "languageCode": "A String", # A user-specified language code for the conversation. + "latestAnalysis": { # The analysis resource. # Output only. The conversation's latest analysis, if one exists. + "analysisResult": { # The result of an analysis. # Output only. The result of the analysis, which is populated when the analysis finishes. + "callAnalysisMetadata": { # Call-specific metadata created during analysis. # Call-specific metadata created by the analysis. + "annotations": [ # A list of call annotations that apply to this call. + { # A piece of metadata that applies to a window of a call. + "annotationEndBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "annotationStartBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "channelTag": 42, # The channel of the audio where the annotation occurs. For single-channel audio, this field is not populated. + "entityMentionData": { # The data for an entity mention annotation. This represents a mention of an `Entity` in the conversation. # Data specifying an entity mention. + "entityUniqueId": "A String", # The key of this entity in conversation entities. Can be used to retrieve the exact `Entity` this mention is attached to. + "sentiment": { # The data for a sentiment annotation. # Sentiment expressed for this mention of the entity. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The type of the entity mention. + }, + "holdData": { # The data for a hold annotation. # Data specifying a hold. + }, + "intentMatchData": { # The data for an intent match. Represents an intent match for a text segment in the conversation. A text segment can be part of a sentence, a complete sentence, or an utterance with multiple sentences. # Data specifying an intent match. + "intentUniqueId": "A String", # The id of the matched intent. Can be used to retrieve the corresponding intent information. + }, + "interruptionData": { # The data for an interruption annotation. # Data specifying an interruption. + }, + "phraseMatchData": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. # Data specifying a phrase match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "silenceData": { # The data for a silence annotation. # Data specifying silence. + }, + }, + ], + "entities": { # All the entities in the call. + "a_key": { # The data for an entity annotation. Represents a phrase in the conversation that is a known entity, such as a person, an organization, or location. + "displayName": "A String", # The representative name for the entity. + "metadata": { # Metadata associated with the entity. For most entity types, the metadata is a Wikipedia URL (`wikipedia_url`) and Knowledge Graph MID (`mid`), if they are available. For the metadata associated with other entity types, see the Type table below. + "a_key": "A String", + }, + "salience": 3.14, # The salience score associated with the entity in the [0, 1.0] range. The salience score for an entity provides information about the importance or centrality of that entity to the entire document text. Scores closer to 0 are less salient, while scores closer to 1.0 are highly salient. + "sentiment": { # The data for a sentiment annotation. # The aggregate sentiment expressed for this entity in the conversation. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The entity type. + }, + }, + "intents": { # All the matched intents in the call. + "a_key": { # The data for an intent. Represents a detected intent in the conversation, for example MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + "id": "A String", # The unique identifier of the intent. + }, + }, + "issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result. + "issueModel": "A String", # Issue model that generates the result. + "issues": [ # All the matched issues. + { # Information about the issue. + "issue": "A String", # Resource name of the assigned issue. + "score": 3.14, # Score indicating the likelihood of the issue assignment. currently bounded on [0,1]. + }, + ], + }, + "phraseMatchers": { # All the matched phrase matchers in the call. + "a_key": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + }, + "sentiments": [ # Overall conversation-level sentiment for each channel of the call. + { # One channel of conversation-level sentiment data. + "channelTag": 42, # The channel of the audio that the data applies to. + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + }, + ], + }, + "endTime": "A String", # The time at which the analysis ended. + }, + "createTime": "A String", # Output only. The time at which the analysis was created, which occurs when the long-running operation completes. + "name": "A String", # Immutable. The resource name of the analysis. Format: projects/{project}/locations/{location}/conversations/{conversation}/analyses/{analysis} + "requestTime": "A String", # Output only. The time at which the analysis was requested. + }, + "medium": "A String", # Immutable. The conversation medium. + "name": "A String", # Immutable. The resource name of the conversation. Format: projects/{project}/locations/{location}/conversations/{conversation} + "runtimeAnnotations": [ # Output only. The annotations that were generated during the customer and agent interaction. + { # An annotation that was generated during the customer and agent interaction. + "annotationId": "A String", # The unique identifier of the annotation. Format: projects/{project}/locations/{location}/conversationDatasets/{dataset}/conversationDataItems/{data_item}/conversationAnnotations/{annotation} + "answerFeedback": { # The feedback that the customer has about a certain answer in the conversation. # The feedback that the customer has about the answer in `data`. + "clicked": True or False, # Indicates whether an answer or item was clicked by the human agent. + "correctnessLevel": "A String", # The correctness level of an answer. + "displayed": True or False, # Indicates whether an answer or item was displayed to the human agent in the agent desktop UI. + }, + "articleSuggestion": { # Agent Assist Article Suggestion data. # Agent Assist Article Suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this article is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Article Suggestion and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document} + "title": "A String", # Article title. + "uri": "A String", # Article URI. + }, + "createTime": "A String", # The time at which this annotation was created. + "dialogflowInteraction": { # Dialogflow interaction data. # Dialogflow interaction data. + "confidence": 3.14, # The confidence of the match ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "dialogflowIntentId": "A String", # The Dialogflow intent resource path. Format: projects/{project}/agent/{agent}/intents/{intent} + }, + "endBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "faqAnswer": { # Agent Assist frequently-asked-question answer data. # Agent Assist FAQ answer data. + "answer": "A String", # The piece of text from the `source` knowledge base document. + "confidenceScore": 3.14, # The system's confidence score that this answer is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the FAQ answer and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record}. + "question": "A String", # The corresponding FAQ question. + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. + }, + "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "suggestion": "A String", # The content of the suggestion. + }, + "smartReply": { # Agent Assist Smart Reply data. # Agent Assist Smart Reply data. + "confidenceScore": 3.14, # The system's confidence score that this reply is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Reply and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "reply": "A String", # The content of the reply. + }, + "startBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + }, + ], + "startTime": "A String", # The time at which the conversation started. + "transcript": { # A message representing the transcript of a conversation. # Output only. The conversation transcript. + "transcriptSegments": [ # A list of sequential transcript segments that comprise the conversation. + { # A segment of a full transcript. + "channelTag": 42, # For conversations derived from multi-channel audio, this is the channel number corresponding to the audio from that channel. For audioChannelCount = N, its output values can range from '1' to 'N'. A channel tag of 0 indicates that the audio is mono. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this segment. A default value of 0.0 indicates that the value is unset. + "languageCode": "A String", # The language code of this segment as a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US". + "segmentParticipant": { # The call participant speaking for a given utterance. # The participant of this segment. + "dialogflowParticipant": "A String", # The name of the Dialogflow participant. Format: projects/{project}/locations/{location}/conversations/{conversation}/participants/{participant} + "role": "A String", # The role of the participant. + }, + "text": "A String", # The text of this segment. + "words": [ # A list of the word-specific information for each word in the segment. + { # Word-level info for words in a transcript. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this word. A default value of 0.0 indicates that the value is unset. + "endOffset": "A String", # Time offset of the end of this word relative to the beginning of the total conversation. + "startOffset": "A String", # Time offset of the start of this word relative to the beginning of the total conversation. + "word": "A String", # The word itself. Includes punctuation marks that surround the word. + }, + ], + }, + ], + }, + "ttl": "A String", # Input only. The TTL for this resource. If specified, then this TTL will be used to calculate the expire time. + "turnCount": 42, # Output only. The number of turns in the conversation. + "updateTime": "A String", # Output only. The most recent time at which the conversation was updated. +}+++ +delete(name, force=None, x__xgafv=None)+Deletes a conversation. + +Args: + name: string, Required. The name of the conversation to delete. (required) + force: boolean, If set to true, all of this conversation's analyses will also be deleted. Otherwise, the request will only succeed if the conversation has no analyses. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`. +}+++ +get(name, view=None, x__xgafv=None)+Gets a conversation. + +Args: + name: string, Required. The name of the conversation to get. (required) + view: string, The level of details of the conversation. Default is `FULL`. + Allowed values + CONVERSATION_VIEW_UNSPECIFIED - Not specified. Defaults to FULL on GetConversationRequest and BASIC for ListConversationsRequest. + BASIC - Transcript field is not populated in the response. + FULL - All fields are populated. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The conversation resource. + "agentId": "A String", # An opaque, user-specified string representing the human agent who handled the conversation. + "callMetadata": { # Call-specific metadata. # Call-specific metadata. + "agentChannel": 42, # The audio channel that contains the agent. + "customerChannel": 42, # The audio channel that contains the customer. + }, + "createTime": "A String", # Output only. The time at which the conversation was created. + "dataSource": { # The conversation source, which is a combination of transcript and audio. # The source of the audio and transcription for the conversation. + "dialogflowSource": { # A Dialogflow source of conversation data. # The source when the conversation comes from Dialogflow. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "dialogflowConversation": "A String", # Output only. The name of the Dialogflow conversation that this conversation resource is derived from. Format: projects/{project}/locations/{location}/conversations/{conversation} + }, + "gcsSource": { # A Cloud Storage source of conversation data. # A Cloud Storage location specification for the audio and transcript. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "transcriptUri": "A String", # Immutable. Cloud Storage URI that points to a file that contains the conversation transcript. + }, + }, + "dialogflowIntents": { # Output only. All the matched Dialogflow intents in the call. The key corresponds to a Dialogflow intent, format: projects/{project}/agent/{agent}/intents/{intent} + "a_key": { # The data for a Dialogflow intent. Represents a detected intent in the conversation, e.g. MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + }, + }, + "duration": "A String", # Output only. The duration of the conversation. + "expireTime": "A String", # The time at which this conversation should expire. After this time, the conversation data and any associated analyses will be deleted. + "labels": { # A map for the user to specify any custom fields. A maximum of 20 labels per conversation is allowed, with a maximum of 256 characters per entry. + "a_key": "A String", + }, + "languageCode": "A String", # A user-specified language code for the conversation. + "latestAnalysis": { # The analysis resource. # Output only. The conversation's latest analysis, if one exists. + "analysisResult": { # The result of an analysis. # Output only. The result of the analysis, which is populated when the analysis finishes. + "callAnalysisMetadata": { # Call-specific metadata created during analysis. # Call-specific metadata created by the analysis. + "annotations": [ # A list of call annotations that apply to this call. + { # A piece of metadata that applies to a window of a call. + "annotationEndBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "annotationStartBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "channelTag": 42, # The channel of the audio where the annotation occurs. For single-channel audio, this field is not populated. + "entityMentionData": { # The data for an entity mention annotation. This represents a mention of an `Entity` in the conversation. # Data specifying an entity mention. + "entityUniqueId": "A String", # The key of this entity in conversation entities. Can be used to retrieve the exact `Entity` this mention is attached to. + "sentiment": { # The data for a sentiment annotation. # Sentiment expressed for this mention of the entity. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The type of the entity mention. + }, + "holdData": { # The data for a hold annotation. # Data specifying a hold. + }, + "intentMatchData": { # The data for an intent match. Represents an intent match for a text segment in the conversation. A text segment can be part of a sentence, a complete sentence, or an utterance with multiple sentences. # Data specifying an intent match. + "intentUniqueId": "A String", # The id of the matched intent. Can be used to retrieve the corresponding intent information. + }, + "interruptionData": { # The data for an interruption annotation. # Data specifying an interruption. + }, + "phraseMatchData": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. # Data specifying a phrase match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "silenceData": { # The data for a silence annotation. # Data specifying silence. + }, + }, + ], + "entities": { # All the entities in the call. + "a_key": { # The data for an entity annotation. Represents a phrase in the conversation that is a known entity, such as a person, an organization, or location. + "displayName": "A String", # The representative name for the entity. + "metadata": { # Metadata associated with the entity. For most entity types, the metadata is a Wikipedia URL (`wikipedia_url`) and Knowledge Graph MID (`mid`), if they are available. For the metadata associated with other entity types, see the Type table below. + "a_key": "A String", + }, + "salience": 3.14, # The salience score associated with the entity in the [0, 1.0] range. The salience score for an entity provides information about the importance or centrality of that entity to the entire document text. Scores closer to 0 are less salient, while scores closer to 1.0 are highly salient. + "sentiment": { # The data for a sentiment annotation. # The aggregate sentiment expressed for this entity in the conversation. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The entity type. + }, + }, + "intents": { # All the matched intents in the call. + "a_key": { # The data for an intent. Represents a detected intent in the conversation, for example MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + "id": "A String", # The unique identifier of the intent. + }, + }, + "issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result. + "issueModel": "A String", # Issue model that generates the result. + "issues": [ # All the matched issues. + { # Information about the issue. + "issue": "A String", # Resource name of the assigned issue. + "score": 3.14, # Score indicating the likelihood of the issue assignment. currently bounded on [0,1]. + }, + ], + }, + "phraseMatchers": { # All the matched phrase matchers in the call. + "a_key": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + }, + "sentiments": [ # Overall conversation-level sentiment for each channel of the call. + { # One channel of conversation-level sentiment data. + "channelTag": 42, # The channel of the audio that the data applies to. + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + }, + ], + }, + "endTime": "A String", # The time at which the analysis ended. + }, + "createTime": "A String", # Output only. The time at which the analysis was created, which occurs when the long-running operation completes. + "name": "A String", # Immutable. The resource name of the analysis. Format: projects/{project}/locations/{location}/conversations/{conversation}/analyses/{analysis} + "requestTime": "A String", # Output only. The time at which the analysis was requested. + }, + "medium": "A String", # Immutable. The conversation medium. + "name": "A String", # Immutable. The resource name of the conversation. Format: projects/{project}/locations/{location}/conversations/{conversation} + "runtimeAnnotations": [ # Output only. The annotations that were generated during the customer and agent interaction. + { # An annotation that was generated during the customer and agent interaction. + "annotationId": "A String", # The unique identifier of the annotation. Format: projects/{project}/locations/{location}/conversationDatasets/{dataset}/conversationDataItems/{data_item}/conversationAnnotations/{annotation} + "answerFeedback": { # The feedback that the customer has about a certain answer in the conversation. # The feedback that the customer has about the answer in `data`. + "clicked": True or False, # Indicates whether an answer or item was clicked by the human agent. + "correctnessLevel": "A String", # The correctness level of an answer. + "displayed": True or False, # Indicates whether an answer or item was displayed to the human agent in the agent desktop UI. + }, + "articleSuggestion": { # Agent Assist Article Suggestion data. # Agent Assist Article Suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this article is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Article Suggestion and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document} + "title": "A String", # Article title. + "uri": "A String", # Article URI. + }, + "createTime": "A String", # The time at which this annotation was created. + "dialogflowInteraction": { # Dialogflow interaction data. # Dialogflow interaction data. + "confidence": 3.14, # The confidence of the match ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "dialogflowIntentId": "A String", # The Dialogflow intent resource path. Format: projects/{project}/agent/{agent}/intents/{intent} + }, + "endBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "faqAnswer": { # Agent Assist frequently-asked-question answer data. # Agent Assist FAQ answer data. + "answer": "A String", # The piece of text from the `source` knowledge base document. + "confidenceScore": 3.14, # The system's confidence score that this answer is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the FAQ answer and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record}. + "question": "A String", # The corresponding FAQ question. + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. + }, + "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "suggestion": "A String", # The content of the suggestion. + }, + "smartReply": { # Agent Assist Smart Reply data. # Agent Assist Smart Reply data. + "confidenceScore": 3.14, # The system's confidence score that this reply is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Reply and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "reply": "A String", # The content of the reply. + }, + "startBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + }, + ], + "startTime": "A String", # The time at which the conversation started. + "transcript": { # A message representing the transcript of a conversation. # Output only. The conversation transcript. + "transcriptSegments": [ # A list of sequential transcript segments that comprise the conversation. + { # A segment of a full transcript. + "channelTag": 42, # For conversations derived from multi-channel audio, this is the channel number corresponding to the audio from that channel. For audioChannelCount = N, its output values can range from '1' to 'N'. A channel tag of 0 indicates that the audio is mono. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this segment. A default value of 0.0 indicates that the value is unset. + "languageCode": "A String", # The language code of this segment as a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US". + "segmentParticipant": { # The call participant speaking for a given utterance. # The participant of this segment. + "dialogflowParticipant": "A String", # The name of the Dialogflow participant. Format: projects/{project}/locations/{location}/conversations/{conversation}/participants/{participant} + "role": "A String", # The role of the participant. + }, + "text": "A String", # The text of this segment. + "words": [ # A list of the word-specific information for each word in the segment. + { # Word-level info for words in a transcript. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this word. A default value of 0.0 indicates that the value is unset. + "endOffset": "A String", # Time offset of the end of this word relative to the beginning of the total conversation. + "startOffset": "A String", # Time offset of the start of this word relative to the beginning of the total conversation. + "word": "A String", # The word itself. Includes punctuation marks that surround the word. + }, + ], + }, + ], + }, + "ttl": "A String", # Input only. The TTL for this resource. If specified, then this TTL will be used to calculate the expire time. + "turnCount": 42, # Output only. The number of turns in the conversation. + "updateTime": "A String", # Output only. The most recent time at which the conversation was updated. +}+++ +list(parent, filter=None, pageSize=None, pageToken=None, view=None, x__xgafv=None)+Lists conversations. + +Args: + parent: string, Required. The parent resource of the conversation. (required) + filter: string, A filter to reduce results to a specific subset. Useful for querying conversations with specific properties. + pageSize: integer, The maximum number of conversations to return in the response. If this value is zero, the service will select a default size. A call might return fewer objects than requested. A non-empty `next_page_token` in the response indicates that more data is available. + pageToken: string, The value returned by the last `ListConversationsResponse`. This value indicates that this is a continuation of a prior `ListConversations` call and that the system should return the next page of data. + view: string, The level of details of the conversation. Default is `BASIC`. + Allowed values + CONVERSATION_VIEW_UNSPECIFIED - Not specified. Defaults to FULL on GetConversationRequest and BASIC for ListConversationsRequest. + BASIC - Transcript field is not populated in the response. + FULL - All fields are populated. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response of listing conversations. + "conversations": [ # The conversations that match the request. + { # The conversation resource. + "agentId": "A String", # An opaque, user-specified string representing the human agent who handled the conversation. + "callMetadata": { # Call-specific metadata. # Call-specific metadata. + "agentChannel": 42, # The audio channel that contains the agent. + "customerChannel": 42, # The audio channel that contains the customer. + }, + "createTime": "A String", # Output only. The time at which the conversation was created. + "dataSource": { # The conversation source, which is a combination of transcript and audio. # The source of the audio and transcription for the conversation. + "dialogflowSource": { # A Dialogflow source of conversation data. # The source when the conversation comes from Dialogflow. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "dialogflowConversation": "A String", # Output only. The name of the Dialogflow conversation that this conversation resource is derived from. Format: projects/{project}/locations/{location}/conversations/{conversation} + }, + "gcsSource": { # A Cloud Storage source of conversation data. # A Cloud Storage location specification for the audio and transcript. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "transcriptUri": "A String", # Immutable. Cloud Storage URI that points to a file that contains the conversation transcript. + }, + }, + "dialogflowIntents": { # Output only. All the matched Dialogflow intents in the call. The key corresponds to a Dialogflow intent, format: projects/{project}/agent/{agent}/intents/{intent} + "a_key": { # The data for a Dialogflow intent. Represents a detected intent in the conversation, e.g. MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + }, + }, + "duration": "A String", # Output only. The duration of the conversation. + "expireTime": "A String", # The time at which this conversation should expire. After this time, the conversation data and any associated analyses will be deleted. + "labels": { # A map for the user to specify any custom fields. A maximum of 20 labels per conversation is allowed, with a maximum of 256 characters per entry. + "a_key": "A String", + }, + "languageCode": "A String", # A user-specified language code for the conversation. + "latestAnalysis": { # The analysis resource. # Output only. The conversation's latest analysis, if one exists. + "analysisResult": { # The result of an analysis. # Output only. The result of the analysis, which is populated when the analysis finishes. + "callAnalysisMetadata": { # Call-specific metadata created during analysis. # Call-specific metadata created by the analysis. + "annotations": [ # A list of call annotations that apply to this call. + { # A piece of metadata that applies to a window of a call. + "annotationEndBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "annotationStartBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "channelTag": 42, # The channel of the audio where the annotation occurs. For single-channel audio, this field is not populated. + "entityMentionData": { # The data for an entity mention annotation. This represents a mention of an `Entity` in the conversation. # Data specifying an entity mention. + "entityUniqueId": "A String", # The key of this entity in conversation entities. Can be used to retrieve the exact `Entity` this mention is attached to. + "sentiment": { # The data for a sentiment annotation. # Sentiment expressed for this mention of the entity. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The type of the entity mention. + }, + "holdData": { # The data for a hold annotation. # Data specifying a hold. + }, + "intentMatchData": { # The data for an intent match. Represents an intent match for a text segment in the conversation. A text segment can be part of a sentence, a complete sentence, or an utterance with multiple sentences. # Data specifying an intent match. + "intentUniqueId": "A String", # The id of the matched intent. Can be used to retrieve the corresponding intent information. + }, + "interruptionData": { # The data for an interruption annotation. # Data specifying an interruption. + }, + "phraseMatchData": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. # Data specifying a phrase match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "silenceData": { # The data for a silence annotation. # Data specifying silence. + }, + }, + ], + "entities": { # All the entities in the call. + "a_key": { # The data for an entity annotation. Represents a phrase in the conversation that is a known entity, such as a person, an organization, or location. + "displayName": "A String", # The representative name for the entity. + "metadata": { # Metadata associated with the entity. For most entity types, the metadata is a Wikipedia URL (`wikipedia_url`) and Knowledge Graph MID (`mid`), if they are available. For the metadata associated with other entity types, see the Type table below. + "a_key": "A String", + }, + "salience": 3.14, # The salience score associated with the entity in the [0, 1.0] range. The salience score for an entity provides information about the importance or centrality of that entity to the entire document text. Scores closer to 0 are less salient, while scores closer to 1.0 are highly salient. + "sentiment": { # The data for a sentiment annotation. # The aggregate sentiment expressed for this entity in the conversation. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The entity type. + }, + }, + "intents": { # All the matched intents in the call. + "a_key": { # The data for an intent. Represents a detected intent in the conversation, for example MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + "id": "A String", # The unique identifier of the intent. + }, + }, + "issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result. + "issueModel": "A String", # Issue model that generates the result. + "issues": [ # All the matched issues. + { # Information about the issue. + "issue": "A String", # Resource name of the assigned issue. + "score": 3.14, # Score indicating the likelihood of the issue assignment. currently bounded on [0,1]. + }, + ], + }, + "phraseMatchers": { # All the matched phrase matchers in the call. + "a_key": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + }, + "sentiments": [ # Overall conversation-level sentiment for each channel of the call. + { # One channel of conversation-level sentiment data. + "channelTag": 42, # The channel of the audio that the data applies to. + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + }, + ], + }, + "endTime": "A String", # The time at which the analysis ended. + }, + "createTime": "A String", # Output only. The time at which the analysis was created, which occurs when the long-running operation completes. + "name": "A String", # Immutable. The resource name of the analysis. Format: projects/{project}/locations/{location}/conversations/{conversation}/analyses/{analysis} + "requestTime": "A String", # Output only. The time at which the analysis was requested. + }, + "medium": "A String", # Immutable. The conversation medium. + "name": "A String", # Immutable. The resource name of the conversation. Format: projects/{project}/locations/{location}/conversations/{conversation} + "runtimeAnnotations": [ # Output only. The annotations that were generated during the customer and agent interaction. + { # An annotation that was generated during the customer and agent interaction. + "annotationId": "A String", # The unique identifier of the annotation. Format: projects/{project}/locations/{location}/conversationDatasets/{dataset}/conversationDataItems/{data_item}/conversationAnnotations/{annotation} + "answerFeedback": { # The feedback that the customer has about a certain answer in the conversation. # The feedback that the customer has about the answer in `data`. + "clicked": True or False, # Indicates whether an answer or item was clicked by the human agent. + "correctnessLevel": "A String", # The correctness level of an answer. + "displayed": True or False, # Indicates whether an answer or item was displayed to the human agent in the agent desktop UI. + }, + "articleSuggestion": { # Agent Assist Article Suggestion data. # Agent Assist Article Suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this article is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Article Suggestion and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document} + "title": "A String", # Article title. + "uri": "A String", # Article URI. + }, + "createTime": "A String", # The time at which this annotation was created. + "dialogflowInteraction": { # Dialogflow interaction data. # Dialogflow interaction data. + "confidence": 3.14, # The confidence of the match ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "dialogflowIntentId": "A String", # The Dialogflow intent resource path. Format: projects/{project}/agent/{agent}/intents/{intent} + }, + "endBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "faqAnswer": { # Agent Assist frequently-asked-question answer data. # Agent Assist FAQ answer data. + "answer": "A String", # The piece of text from the `source` knowledge base document. + "confidenceScore": 3.14, # The system's confidence score that this answer is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the FAQ answer and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record}. + "question": "A String", # The corresponding FAQ question. + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. + }, + "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "suggestion": "A String", # The content of the suggestion. + }, + "smartReply": { # Agent Assist Smart Reply data. # Agent Assist Smart Reply data. + "confidenceScore": 3.14, # The system's confidence score that this reply is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Reply and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "reply": "A String", # The content of the reply. + }, + "startBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + }, + ], + "startTime": "A String", # The time at which the conversation started. + "transcript": { # A message representing the transcript of a conversation. # Output only. The conversation transcript. + "transcriptSegments": [ # A list of sequential transcript segments that comprise the conversation. + { # A segment of a full transcript. + "channelTag": 42, # For conversations derived from multi-channel audio, this is the channel number corresponding to the audio from that channel. For audioChannelCount = N, its output values can range from '1' to 'N'. A channel tag of 0 indicates that the audio is mono. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this segment. A default value of 0.0 indicates that the value is unset. + "languageCode": "A String", # The language code of this segment as a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US". + "segmentParticipant": { # The call participant speaking for a given utterance. # The participant of this segment. + "dialogflowParticipant": "A String", # The name of the Dialogflow participant. Format: projects/{project}/locations/{location}/conversations/{conversation}/participants/{participant} + "role": "A String", # The role of the participant. + }, + "text": "A String", # The text of this segment. + "words": [ # A list of the word-specific information for each word in the segment. + { # Word-level info for words in a transcript. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this word. A default value of 0.0 indicates that the value is unset. + "endOffset": "A String", # Time offset of the end of this word relative to the beginning of the total conversation. + "startOffset": "A String", # Time offset of the start of this word relative to the beginning of the total conversation. + "word": "A String", # The word itself. Includes punctuation marks that surround the word. + }, + ], + }, + ], + }, + "ttl": "A String", # Input only. The TTL for this resource. If specified, then this TTL will be used to calculate the expire time. + "turnCount": 42, # Output only. The number of turns in the conversation. + "updateTime": "A String", # Output only. The most recent time at which the conversation was updated. + }, + ], + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. +}+++ +list_next(previous_request, previous_response)+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++++ + \ No newline at end of file diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.html b/docs/dyn/contactcenterinsights_v1.projects.locations.html new file mode 100644 index 00000000000..a093cd3b43f --- /dev/null +++ b/docs/dyn/contactcenterinsights_v1.projects.locations.html @@ -0,0 +1,193 @@ + + + +patch(name, body=None, updateMask=None, x__xgafv=None)+Updates a conversation. + +Args: + name: string, Immutable. The resource name of the conversation. Format: projects/{project}/locations/{location}/conversations/{conversation} (required) + body: object, The request body. + The object takes the form of: + +{ # The conversation resource. + "agentId": "A String", # An opaque, user-specified string representing the human agent who handled the conversation. + "callMetadata": { # Call-specific metadata. # Call-specific metadata. + "agentChannel": 42, # The audio channel that contains the agent. + "customerChannel": 42, # The audio channel that contains the customer. + }, + "createTime": "A String", # Output only. The time at which the conversation was created. + "dataSource": { # The conversation source, which is a combination of transcript and audio. # The source of the audio and transcription for the conversation. + "dialogflowSource": { # A Dialogflow source of conversation data. # The source when the conversation comes from Dialogflow. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "dialogflowConversation": "A String", # Output only. The name of the Dialogflow conversation that this conversation resource is derived from. Format: projects/{project}/locations/{location}/conversations/{conversation} + }, + "gcsSource": { # A Cloud Storage source of conversation data. # A Cloud Storage location specification for the audio and transcript. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "transcriptUri": "A String", # Immutable. Cloud Storage URI that points to a file that contains the conversation transcript. + }, + }, + "dialogflowIntents": { # Output only. All the matched Dialogflow intents in the call. The key corresponds to a Dialogflow intent, format: projects/{project}/agent/{agent}/intents/{intent} + "a_key": { # The data for a Dialogflow intent. Represents a detected intent in the conversation, e.g. MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + }, + }, + "duration": "A String", # Output only. The duration of the conversation. + "expireTime": "A String", # The time at which this conversation should expire. After this time, the conversation data and any associated analyses will be deleted. + "labels": { # A map for the user to specify any custom fields. A maximum of 20 labels per conversation is allowed, with a maximum of 256 characters per entry. + "a_key": "A String", + }, + "languageCode": "A String", # A user-specified language code for the conversation. + "latestAnalysis": { # The analysis resource. # Output only. The conversation's latest analysis, if one exists. + "analysisResult": { # The result of an analysis. # Output only. The result of the analysis, which is populated when the analysis finishes. + "callAnalysisMetadata": { # Call-specific metadata created during analysis. # Call-specific metadata created by the analysis. + "annotations": [ # A list of call annotations that apply to this call. + { # A piece of metadata that applies to a window of a call. + "annotationEndBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "annotationStartBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "channelTag": 42, # The channel of the audio where the annotation occurs. For single-channel audio, this field is not populated. + "entityMentionData": { # The data for an entity mention annotation. This represents a mention of an `Entity` in the conversation. # Data specifying an entity mention. + "entityUniqueId": "A String", # The key of this entity in conversation entities. Can be used to retrieve the exact `Entity` this mention is attached to. + "sentiment": { # The data for a sentiment annotation. # Sentiment expressed for this mention of the entity. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The type of the entity mention. + }, + "holdData": { # The data for a hold annotation. # Data specifying a hold. + }, + "intentMatchData": { # The data for an intent match. Represents an intent match for a text segment in the conversation. A text segment can be part of a sentence, a complete sentence, or an utterance with multiple sentences. # Data specifying an intent match. + "intentUniqueId": "A String", # The id of the matched intent. Can be used to retrieve the corresponding intent information. + }, + "interruptionData": { # The data for an interruption annotation. # Data specifying an interruption. + }, + "phraseMatchData": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. # Data specifying a phrase match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "silenceData": { # The data for a silence annotation. # Data specifying silence. + }, + }, + ], + "entities": { # All the entities in the call. + "a_key": { # The data for an entity annotation. Represents a phrase in the conversation that is a known entity, such as a person, an organization, or location. + "displayName": "A String", # The representative name for the entity. + "metadata": { # Metadata associated with the entity. For most entity types, the metadata is a Wikipedia URL (`wikipedia_url`) and Knowledge Graph MID (`mid`), if they are available. For the metadata associated with other entity types, see the Type table below. + "a_key": "A String", + }, + "salience": 3.14, # The salience score associated with the entity in the [0, 1.0] range. The salience score for an entity provides information about the importance or centrality of that entity to the entire document text. Scores closer to 0 are less salient, while scores closer to 1.0 are highly salient. + "sentiment": { # The data for a sentiment annotation. # The aggregate sentiment expressed for this entity in the conversation. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The entity type. + }, + }, + "intents": { # All the matched intents in the call. + "a_key": { # The data for an intent. Represents a detected intent in the conversation, for example MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + "id": "A String", # The unique identifier of the intent. + }, + }, + "issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result. + "issueModel": "A String", # Issue model that generates the result. + "issues": [ # All the matched issues. + { # Information about the issue. + "issue": "A String", # Resource name of the assigned issue. + "score": 3.14, # Score indicating the likelihood of the issue assignment. currently bounded on [0,1]. + }, + ], + }, + "phraseMatchers": { # All the matched phrase matchers in the call. + "a_key": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + }, + "sentiments": [ # Overall conversation-level sentiment for each channel of the call. + { # One channel of conversation-level sentiment data. + "channelTag": 42, # The channel of the audio that the data applies to. + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + }, + ], + }, + "endTime": "A String", # The time at which the analysis ended. + }, + "createTime": "A String", # Output only. The time at which the analysis was created, which occurs when the long-running operation completes. + "name": "A String", # Immutable. The resource name of the analysis. Format: projects/{project}/locations/{location}/conversations/{conversation}/analyses/{analysis} + "requestTime": "A String", # Output only. The time at which the analysis was requested. + }, + "medium": "A String", # Immutable. The conversation medium. + "name": "A String", # Immutable. The resource name of the conversation. Format: projects/{project}/locations/{location}/conversations/{conversation} + "runtimeAnnotations": [ # Output only. The annotations that were generated during the customer and agent interaction. + { # An annotation that was generated during the customer and agent interaction. + "annotationId": "A String", # The unique identifier of the annotation. Format: projects/{project}/locations/{location}/conversationDatasets/{dataset}/conversationDataItems/{data_item}/conversationAnnotations/{annotation} + "answerFeedback": { # The feedback that the customer has about a certain answer in the conversation. # The feedback that the customer has about the answer in `data`. + "clicked": True or False, # Indicates whether an answer or item was clicked by the human agent. + "correctnessLevel": "A String", # The correctness level of an answer. + "displayed": True or False, # Indicates whether an answer or item was displayed to the human agent in the agent desktop UI. + }, + "articleSuggestion": { # Agent Assist Article Suggestion data. # Agent Assist Article Suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this article is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Article Suggestion and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document} + "title": "A String", # Article title. + "uri": "A String", # Article URI. + }, + "createTime": "A String", # The time at which this annotation was created. + "dialogflowInteraction": { # Dialogflow interaction data. # Dialogflow interaction data. + "confidence": 3.14, # The confidence of the match ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "dialogflowIntentId": "A String", # The Dialogflow intent resource path. Format: projects/{project}/agent/{agent}/intents/{intent} + }, + "endBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "faqAnswer": { # Agent Assist frequently-asked-question answer data. # Agent Assist FAQ answer data. + "answer": "A String", # The piece of text from the `source` knowledge base document. + "confidenceScore": 3.14, # The system's confidence score that this answer is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the FAQ answer and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record}. + "question": "A String", # The corresponding FAQ question. + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. + }, + "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "suggestion": "A String", # The content of the suggestion. + }, + "smartReply": { # Agent Assist Smart Reply data. # Agent Assist Smart Reply data. + "confidenceScore": 3.14, # The system's confidence score that this reply is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Reply and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "reply": "A String", # The content of the reply. + }, + "startBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + }, + ], + "startTime": "A String", # The time at which the conversation started. + "transcript": { # A message representing the transcript of a conversation. # Output only. The conversation transcript. + "transcriptSegments": [ # A list of sequential transcript segments that comprise the conversation. + { # A segment of a full transcript. + "channelTag": 42, # For conversations derived from multi-channel audio, this is the channel number corresponding to the audio from that channel. For audioChannelCount = N, its output values can range from '1' to 'N'. A channel tag of 0 indicates that the audio is mono. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this segment. A default value of 0.0 indicates that the value is unset. + "languageCode": "A String", # The language code of this segment as a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US". + "segmentParticipant": { # The call participant speaking for a given utterance. # The participant of this segment. + "dialogflowParticipant": "A String", # The name of the Dialogflow participant. Format: projects/{project}/locations/{location}/conversations/{conversation}/participants/{participant} + "role": "A String", # The role of the participant. + }, + "text": "A String", # The text of this segment. + "words": [ # A list of the word-specific information for each word in the segment. + { # Word-level info for words in a transcript. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this word. A default value of 0.0 indicates that the value is unset. + "endOffset": "A String", # Time offset of the end of this word relative to the beginning of the total conversation. + "startOffset": "A String", # Time offset of the start of this word relative to the beginning of the total conversation. + "word": "A String", # The word itself. Includes punctuation marks that surround the word. + }, + ], + }, + ], + }, + "ttl": "A String", # Input only. The TTL for this resource. If specified, then this TTL will be used to calculate the expire time. + "turnCount": 42, # Output only. The number of turns in the conversation. + "updateTime": "A String", # Output only. The most recent time at which the conversation was updated. +} + + updateMask: string, The list of fields to be updated. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The conversation resource. + "agentId": "A String", # An opaque, user-specified string representing the human agent who handled the conversation. + "callMetadata": { # Call-specific metadata. # Call-specific metadata. + "agentChannel": 42, # The audio channel that contains the agent. + "customerChannel": 42, # The audio channel that contains the customer. + }, + "createTime": "A String", # Output only. The time at which the conversation was created. + "dataSource": { # The conversation source, which is a combination of transcript and audio. # The source of the audio and transcription for the conversation. + "dialogflowSource": { # A Dialogflow source of conversation data. # The source when the conversation comes from Dialogflow. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "dialogflowConversation": "A String", # Output only. The name of the Dialogflow conversation that this conversation resource is derived from. Format: projects/{project}/locations/{location}/conversations/{conversation} + }, + "gcsSource": { # A Cloud Storage source of conversation data. # A Cloud Storage location specification for the audio and transcript. + "audioUri": "A String", # Cloud Storage URI that points to a file that contains the conversation audio. + "transcriptUri": "A String", # Immutable. Cloud Storage URI that points to a file that contains the conversation transcript. + }, + }, + "dialogflowIntents": { # Output only. All the matched Dialogflow intents in the call. The key corresponds to a Dialogflow intent, format: projects/{project}/agent/{agent}/intents/{intent} + "a_key": { # The data for a Dialogflow intent. Represents a detected intent in the conversation, e.g. MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + }, + }, + "duration": "A String", # Output only. The duration of the conversation. + "expireTime": "A String", # The time at which this conversation should expire. After this time, the conversation data and any associated analyses will be deleted. + "labels": { # A map for the user to specify any custom fields. A maximum of 20 labels per conversation is allowed, with a maximum of 256 characters per entry. + "a_key": "A String", + }, + "languageCode": "A String", # A user-specified language code for the conversation. + "latestAnalysis": { # The analysis resource. # Output only. The conversation's latest analysis, if one exists. + "analysisResult": { # The result of an analysis. # Output only. The result of the analysis, which is populated when the analysis finishes. + "callAnalysisMetadata": { # Call-specific metadata created during analysis. # Call-specific metadata created by the analysis. + "annotations": [ # A list of call annotations that apply to this call. + { # A piece of metadata that applies to a window of a call. + "annotationEndBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "annotationStartBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "channelTag": 42, # The channel of the audio where the annotation occurs. For single-channel audio, this field is not populated. + "entityMentionData": { # The data for an entity mention annotation. This represents a mention of an `Entity` in the conversation. # Data specifying an entity mention. + "entityUniqueId": "A String", # The key of this entity in conversation entities. Can be used to retrieve the exact `Entity` this mention is attached to. + "sentiment": { # The data for a sentiment annotation. # Sentiment expressed for this mention of the entity. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The type of the entity mention. + }, + "holdData": { # The data for a hold annotation. # Data specifying a hold. + }, + "intentMatchData": { # The data for an intent match. Represents an intent match for a text segment in the conversation. A text segment can be part of a sentence, a complete sentence, or an utterance with multiple sentences. # Data specifying an intent match. + "intentUniqueId": "A String", # The id of the matched intent. Can be used to retrieve the corresponding intent information. + }, + "interruptionData": { # The data for an interruption annotation. # Data specifying an interruption. + }, + "phraseMatchData": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. # Data specifying a phrase match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "silenceData": { # The data for a silence annotation. # Data specifying silence. + }, + }, + ], + "entities": { # All the entities in the call. + "a_key": { # The data for an entity annotation. Represents a phrase in the conversation that is a known entity, such as a person, an organization, or location. + "displayName": "A String", # The representative name for the entity. + "metadata": { # Metadata associated with the entity. For most entity types, the metadata is a Wikipedia URL (`wikipedia_url`) and Knowledge Graph MID (`mid`), if they are available. For the metadata associated with other entity types, see the Type table below. + "a_key": "A String", + }, + "salience": 3.14, # The salience score associated with the entity in the [0, 1.0] range. The salience score for an entity provides information about the importance or centrality of that entity to the entire document text. Scores closer to 0 are less salient, while scores closer to 1.0 are highly salient. + "sentiment": { # The data for a sentiment annotation. # The aggregate sentiment expressed for this entity in the conversation. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + "type": "A String", # The entity type. + }, + }, + "intents": { # All the matched intents in the call. + "a_key": { # The data for an intent. Represents a detected intent in the conversation, for example MAKES_PROMISE. + "displayName": "A String", # The human-readable name of the intent. + "id": "A String", # The unique identifier of the intent. + }, + }, + "issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result. + "issueModel": "A String", # Issue model that generates the result. + "issues": [ # All the matched issues. + { # Information about the issue. + "issue": "A String", # Resource name of the assigned issue. + "score": 3.14, # Score indicating the likelihood of the issue assignment. currently bounded on [0,1]. + }, + ], + }, + "phraseMatchers": { # All the matched phrase matchers in the call. + "a_key": { # The data for a matched phrase matcher. Represents information identifying a phrase matcher for a given match. + "displayName": "A String", # The human-readable name of the phrase matcher. + "phraseMatcher": "A String", # The unique identifier (the resource name) of the phrase matcher. + }, + }, + "sentiments": [ # Overall conversation-level sentiment for each channel of the call. + { # One channel of conversation-level sentiment data. + "channelTag": 42, # The channel of the audio that the data applies to. + "sentimentData": { # The data for a sentiment annotation. # Data specifying sentiment. + "magnitude": 3.14, # A non-negative number from 0 to infinity which represents the abolute magnitude of sentiment regardless of score. + "score": 3.14, # The sentiment score between -1.0 (negative) and 1.0 (positive). + }, + }, + ], + }, + "endTime": "A String", # The time at which the analysis ended. + }, + "createTime": "A String", # Output only. The time at which the analysis was created, which occurs when the long-running operation completes. + "name": "A String", # Immutable. The resource name of the analysis. Format: projects/{project}/locations/{location}/conversations/{conversation}/analyses/{analysis} + "requestTime": "A String", # Output only. The time at which the analysis was requested. + }, + "medium": "A String", # Immutable. The conversation medium. + "name": "A String", # Immutable. The resource name of the conversation. Format: projects/{project}/locations/{location}/conversations/{conversation} + "runtimeAnnotations": [ # Output only. The annotations that were generated during the customer and agent interaction. + { # An annotation that was generated during the customer and agent interaction. + "annotationId": "A String", # The unique identifier of the annotation. Format: projects/{project}/locations/{location}/conversationDatasets/{dataset}/conversationDataItems/{data_item}/conversationAnnotations/{annotation} + "answerFeedback": { # The feedback that the customer has about a certain answer in the conversation. # The feedback that the customer has about the answer in `data`. + "clicked": True or False, # Indicates whether an answer or item was clicked by the human agent. + "correctnessLevel": "A String", # The correctness level of an answer. + "displayed": True or False, # Indicates whether an answer or item was displayed to the human agent in the agent desktop UI. + }, + "articleSuggestion": { # Agent Assist Article Suggestion data. # Agent Assist Article Suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this article is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Article Suggestion and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document} + "title": "A String", # Article title. + "uri": "A String", # Article URI. + }, + "createTime": "A String", # The time at which this annotation was created. + "dialogflowInteraction": { # Dialogflow interaction data. # Dialogflow interaction data. + "confidence": 3.14, # The confidence of the match ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "dialogflowIntentId": "A String", # The Dialogflow intent resource path. Format: projects/{project}/agent/{agent}/intents/{intent} + }, + "endBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation ends, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + "faqAnswer": { # Agent Assist frequently-asked-question answer data. # Agent Assist FAQ answer data. + "answer": "A String", # The piece of text from the `source` knowledge base document. + "confidenceScore": 3.14, # The system's confidence score that this answer is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the FAQ answer and the document that it originates from. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record}. + "question": "A String", # The corresponding FAQ question. + "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. + }, + "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. + "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "suggestion": "A String", # The content of the suggestion. + }, + "smartReply": { # Agent Assist Smart Reply data. # Agent Assist Smart Reply data. + "confidenceScore": 3.14, # The system's confidence score that this reply is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # Map that contains metadata about the Smart Reply and the document from which it originates. + "a_key": "A String", + }, + "queryRecord": "A String", # Name of the query record. Format: projects/{project}/locations/{location}/queryRecords/{query_record} + "reply": "A String", # The content of the reply. + }, + "startBoundary": { # A point in a conversation that marks the start or the end of an annotation. # The boundary in the conversation where the annotation starts, inclusive. + "transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. + "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. + }, + }, + ], + "startTime": "A String", # The time at which the conversation started. + "transcript": { # A message representing the transcript of a conversation. # Output only. The conversation transcript. + "transcriptSegments": [ # A list of sequential transcript segments that comprise the conversation. + { # A segment of a full transcript. + "channelTag": 42, # For conversations derived from multi-channel audio, this is the channel number corresponding to the audio from that channel. For audioChannelCount = N, its output values can range from '1' to 'N'. A channel tag of 0 indicates that the audio is mono. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this segment. A default value of 0.0 indicates that the value is unset. + "languageCode": "A String", # The language code of this segment as a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US". + "segmentParticipant": { # The call participant speaking for a given utterance. # The participant of this segment. + "dialogflowParticipant": "A String", # The name of the Dialogflow participant. Format: projects/{project}/locations/{location}/conversations/{conversation}/participants/{participant} + "role": "A String", # The role of the participant. + }, + "text": "A String", # The text of this segment. + "words": [ # A list of the word-specific information for each word in the segment. + { # Word-level info for words in a transcript. + "confidence": 3.14, # A confidence estimate between 0.0 and 1.0 of the fidelity of this word. A default value of 0.0 indicates that the value is unset. + "endOffset": "A String", # Time offset of the end of this word relative to the beginning of the total conversation. + "startOffset": "A String", # Time offset of the start of this word relative to the beginning of the total conversation. + "word": "A String", # The word itself. Includes punctuation marks that surround the word. + }, + ], + }, + ], + }, + "ttl": "A String", # Input only. The TTL for this resource. If specified, then this TTL will be used to calculate the expire time. + "turnCount": 42, # Output only. The number of turns in the conversation. + "updateTime": "A String", # Output only. The most recent time at which the conversation was updated. +}+Contact Center AI Insights API . projects . locations
+Instance Methods
++
+conversations()+Returns the conversations Resource.
+ ++
+insightsdata()+Returns the insightsdata Resource.
+ ++
+issueModels()+Returns the issueModels Resource.
+ ++
+operations()+Returns the operations Resource.
+ ++
+phraseMatchers()+Returns the phraseMatchers Resource.
+ ++
+close()Close httplib2 connections.
++
+getSettings(name, x__xgafv=None)Gets project-level settings.
++
+updateSettings(name, body=None, updateMask=None, x__xgafv=None)Updates project-level settings.
+Method Details
+++ +close()+Close httplib2 connections.+++ +getSettings(name, x__xgafv=None)+Gets project-level settings. + +Args: + name: string, Required. The name of the settings resource to get. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The settings resource. + "analysisConfig": { # Default configuration when creating Analyses in Insights. # Default analysis settings. + "runtimeIntegrationAnalysisPercentage": 3.14, # Percentage of conversations created using Dialogflow runtime integration to analyze automatically, between [0, 100]. + }, + "conversationTtl": "A String", # The default TTL for newly-created conversations. If a conversation has a specified expiration, that value will be used instead. Changing this value will not change the expiration of existing conversations. Conversations with no expire time persist until they are deleted. + "createTime": "A String", # Output only. The time at which the settings was created. + "languageCode": "A String", # A language code to be applied to each transcript segment unless the segment already specifies a language code. Language code defaults to "en-US" if it is neither specified on the segment nor here. + "name": "A String", # Immutable. The resource name of the settings resource. Format: projects/{project}/locations/{location}/settings + "pubsubNotificationSettings": { # A map that maps a notification trigger to a Pub/Sub topic. Each time a specified trigger occurs, Insights will notify the corresponding Pub/Sub topic. Keys are notification triggers. Supported keys are: * "all-triggers": Notify each time any of the supported triggers occurs. * "create-analysis": Notify each time an analysis is created. * "create-conversation": Notify each time a conversation is created. * "export-insights-data": Notify each time an export is complete. Values are Pub/Sub topics. The format of each Pub/Sub topic is: projects/{project}/topics/{topic} + "a_key": "A String", + }, + "updateTime": "A String", # Output only. The time at which the settings were last updated. +}+++ + \ No newline at end of file diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.insightsdata.html b/docs/dyn/contactcenterinsights_v1.projects.locations.insightsdata.html new file mode 100644 index 00000000000..773f26e9216 --- /dev/null +++ b/docs/dyn/contactcenterinsights_v1.projects.locations.insightsdata.html @@ -0,0 +1,138 @@ + + + +updateSettings(name, body=None, updateMask=None, x__xgafv=None)+Updates project-level settings. + +Args: + name: string, Immutable. The resource name of the settings resource. Format: projects/{project}/locations/{location}/settings (required) + body: object, The request body. + The object takes the form of: + +{ # The settings resource. + "analysisConfig": { # Default configuration when creating Analyses in Insights. # Default analysis settings. + "runtimeIntegrationAnalysisPercentage": 3.14, # Percentage of conversations created using Dialogflow runtime integration to analyze automatically, between [0, 100]. + }, + "conversationTtl": "A String", # The default TTL for newly-created conversations. If a conversation has a specified expiration, that value will be used instead. Changing this value will not change the expiration of existing conversations. Conversations with no expire time persist until they are deleted. + "createTime": "A String", # Output only. The time at which the settings was created. + "languageCode": "A String", # A language code to be applied to each transcript segment unless the segment already specifies a language code. Language code defaults to "en-US" if it is neither specified on the segment nor here. + "name": "A String", # Immutable. The resource name of the settings resource. Format: projects/{project}/locations/{location}/settings + "pubsubNotificationSettings": { # A map that maps a notification trigger to a Pub/Sub topic. Each time a specified trigger occurs, Insights will notify the corresponding Pub/Sub topic. Keys are notification triggers. Supported keys are: * "all-triggers": Notify each time any of the supported triggers occurs. * "create-analysis": Notify each time an analysis is created. * "create-conversation": Notify each time a conversation is created. * "export-insights-data": Notify each time an export is complete. Values are Pub/Sub topics. The format of each Pub/Sub topic is: projects/{project}/topics/{topic} + "a_key": "A String", + }, + "updateTime": "A String", # Output only. The time at which the settings were last updated. +} + + updateMask: string, Required. The list of fields to be updated. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The settings resource. + "analysisConfig": { # Default configuration when creating Analyses in Insights. # Default analysis settings. + "runtimeIntegrationAnalysisPercentage": 3.14, # Percentage of conversations created using Dialogflow runtime integration to analyze automatically, between [0, 100]. + }, + "conversationTtl": "A String", # The default TTL for newly-created conversations. If a conversation has a specified expiration, that value will be used instead. Changing this value will not change the expiration of existing conversations. Conversations with no expire time persist until they are deleted. + "createTime": "A String", # Output only. The time at which the settings was created. + "languageCode": "A String", # A language code to be applied to each transcript segment unless the segment already specifies a language code. Language code defaults to "en-US" if it is neither specified on the segment nor here. + "name": "A String", # Immutable. The resource name of the settings resource. Format: projects/{project}/locations/{location}/settings + "pubsubNotificationSettings": { # A map that maps a notification trigger to a Pub/Sub topic. Each time a specified trigger occurs, Insights will notify the corresponding Pub/Sub topic. Keys are notification triggers. Supported keys are: * "all-triggers": Notify each time any of the supported triggers occurs. * "create-analysis": Notify each time an analysis is created. * "create-conversation": Notify each time a conversation is created. * "export-insights-data": Notify each time an export is complete. Values are Pub/Sub topics. The format of each Pub/Sub topic is: projects/{project}/topics/{topic} + "a_key": "A String", + }, + "updateTime": "A String", # Output only. The time at which the settings were last updated. +}+Contact Center AI Insights API . projects . locations . insightsdata
+Instance Methods
++
+close()Close httplib2 connections.
++
+export(parent, body=None, x__xgafv=None)Export insights data to a destination defined in the request body.
+Method Details
+++ +close()+Close httplib2 connections.+++ + \ No newline at end of file diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.issueModels.html b/docs/dyn/contactcenterinsights_v1.projects.locations.issueModels.html new file mode 100644 index 00000000000..2ef45d3ab6d --- /dev/null +++ b/docs/dyn/contactcenterinsights_v1.projects.locations.issueModels.html @@ -0,0 +1,463 @@ + + + +export(parent, body=None, x__xgafv=None)+Export insights data to a destination defined in the request body. + +Args: + parent: string, Required. The parent resource to export data from. (required) + body: object, The request body. + The object takes the form of: + +{ # The request to export insights. + "bigQueryDestination": { # A BigQuery Table Reference. # Specified if sink is a BigQuery table. + "dataset": "A String", # Required. The name of the BigQuery dataset that the snapshot result should be exported to. If this dataset does not exist, the export call returns an INVALID_ARGUMENT error. + "projectId": "A String", # A project ID or number. If specified, then export will attempt to write data to this project instead of the resource project. Otherwise, the resource project will be used. + "table": "A String", # The BigQuery table name to which the insights data should be written. If this table does not exist, the export call returns an INVALID_ARGUMENT error. + }, + "filter": "A String", # A filter to reduce results to a specific subset. Useful for exporting conversations with specific properties. + "kmsKey": "A String", # A fully qualified KMS key name for BigQuery tables protected by CMEK. Format: projects/{project}/locations/{location}/keyRings/{keyring}/cryptoKeys/{key}/cryptoKeyVersions/{version} + "parent": "A String", # Required. The parent resource to export data from. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+Contact Center AI Insights API . projects . locations . issueModels
+Instance Methods
++
+issues()+Returns the issues Resource.
+ ++
+calculateIssueModelStats(issueModel, x__xgafv=None)Gets an issue model's statistics.
++
+close()Close httplib2 connections.
++
+create(parent, body=None, x__xgafv=None)Creates an issue model.
+ +Deletes an issue model.
++
+deploy(name, body=None, x__xgafv=None)Deploys an issue model. Returns an error if a model is already deployed. An issue model can only be used in analysis after it has been deployed.
+ +Gets an issue model.
+ +Lists issue models.
++
+patch(name, body=None, updateMask=None, x__xgafv=None)Updates an issue model.
++
+undeploy(name, body=None, x__xgafv=None)Undeploys an issue model. An issue model can not be used in analysis after it has been undeployed.
+Method Details
+++ +calculateIssueModelStats(issueModel, x__xgafv=None)+Gets an issue model's statistics. + +Args: + issueModel: string, Required. The resource name of the issue model to query against. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response of querying an issue model's statistics. + "currentStats": { # Aggregated statistics about an issue model. # The latest label statistics for the queried issue model. Includes results on both training data and data labeled after deployment. + "analyzedConversationsCount": "A String", # Number of conversations the issue model has analyzed at this point in time. + "issueStats": { # Statistics on each issue. Key is the issue's resource name. + "a_key": { # Aggregated statistics about an issue. + "issue": "A String", # Issue resource. Format: projects/{project}/locations/{location}/issueModels/{issue_model}/issues/{issue} + "labeledConversationsCount": "A String", # Number of conversations attached to the issue at this point in time. + }, + }, + "unclassifiedConversationsCount": "A String", # Number of analyzed conversations for which no issue was applicable at this point in time. + }, +}+++ +close()+Close httplib2 connections.+++ +create(parent, body=None, x__xgafv=None)+Creates an issue model. + +Args: + parent: string, Required. The parent resource of the issue model. (required) + body: object, The request body. + The object takes the form of: + +{ # The issue model resource. + "createTime": "A String", # Output only. The time at which this issue model was created. + "displayName": "A String", # The representative name for the issue model. + "inputDataConfig": { # Configs for the input data used to create the issue model. # Configs for the input data that used to create the issue model. + "medium": "A String", # Required. Medium of conversations used in training data. + "trainingConversationsCount": "A String", # Output only. Number of conversations used in training. Output only. + }, + "name": "A String", # Immutable. The resource name of the issue model. Format: projects/{project}/locations/{location}/issueModels/{issue_model} + "state": "A String", # Output only. State of the model. + "trainingStats": { # Aggregated statistics about an issue model. # Output only. Immutable. The issue model's label statistics on its training data. + "analyzedConversationsCount": "A String", # Number of conversations the issue model has analyzed at this point in time. + "issueStats": { # Statistics on each issue. Key is the issue's resource name. + "a_key": { # Aggregated statistics about an issue. + "issue": "A String", # Issue resource. Format: projects/{project}/locations/{location}/issueModels/{issue_model}/issues/{issue} + "labeledConversationsCount": "A String", # Number of conversations attached to the issue at this point in time. + }, + }, + "unclassifiedConversationsCount": "A String", # Number of analyzed conversations for which no issue was applicable at this point in time. + }, + "updateTime": "A String", # Output only. The most recent time at which the issue model was updated. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +delete(name, x__xgafv=None)+Deletes an issue model. + +Args: + name: string, Required. The name of the issue model to delete. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +deploy(name, body=None, x__xgafv=None)+Deploys an issue model. Returns an error if a model is already deployed. An issue model can only be used in analysis after it has been deployed. + +Args: + name: string, Required. The issue model to deploy. (required) + body: object, The request body. + The object takes the form of: + +{ # The request to deploy an issue model. + "name": "A String", # Required. The issue model to deploy. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +get(name, x__xgafv=None)+Gets an issue model. + +Args: + name: string, Required. The name of the issue model to get. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The issue model resource. + "createTime": "A String", # Output only. The time at which this issue model was created. + "displayName": "A String", # The representative name for the issue model. + "inputDataConfig": { # Configs for the input data used to create the issue model. # Configs for the input data that used to create the issue model. + "medium": "A String", # Required. Medium of conversations used in training data. + "trainingConversationsCount": "A String", # Output only. Number of conversations used in training. Output only. + }, + "name": "A String", # Immutable. The resource name of the issue model. Format: projects/{project}/locations/{location}/issueModels/{issue_model} + "state": "A String", # Output only. State of the model. + "trainingStats": { # Aggregated statistics about an issue model. # Output only. Immutable. The issue model's label statistics on its training data. + "analyzedConversationsCount": "A String", # Number of conversations the issue model has analyzed at this point in time. + "issueStats": { # Statistics on each issue. Key is the issue's resource name. + "a_key": { # Aggregated statistics about an issue. + "issue": "A String", # Issue resource. Format: projects/{project}/locations/{location}/issueModels/{issue_model}/issues/{issue} + "labeledConversationsCount": "A String", # Number of conversations attached to the issue at this point in time. + }, + }, + "unclassifiedConversationsCount": "A String", # Number of analyzed conversations for which no issue was applicable at this point in time. + }, + "updateTime": "A String", # Output only. The most recent time at which the issue model was updated. +}+++ +list(parent, x__xgafv=None)+Lists issue models. + +Args: + parent: string, Required. The parent resource of the issue model. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response of listing issue models. + "issueModels": [ # The issue models that match the request. + { # The issue model resource. + "createTime": "A String", # Output only. The time at which this issue model was created. + "displayName": "A String", # The representative name for the issue model. + "inputDataConfig": { # Configs for the input data used to create the issue model. # Configs for the input data that used to create the issue model. + "medium": "A String", # Required. Medium of conversations used in training data. + "trainingConversationsCount": "A String", # Output only. Number of conversations used in training. Output only. + }, + "name": "A String", # Immutable. The resource name of the issue model. Format: projects/{project}/locations/{location}/issueModels/{issue_model} + "state": "A String", # Output only. State of the model. + "trainingStats": { # Aggregated statistics about an issue model. # Output only. Immutable. The issue model's label statistics on its training data. + "analyzedConversationsCount": "A String", # Number of conversations the issue model has analyzed at this point in time. + "issueStats": { # Statistics on each issue. Key is the issue's resource name. + "a_key": { # Aggregated statistics about an issue. + "issue": "A String", # Issue resource. Format: projects/{project}/locations/{location}/issueModels/{issue_model}/issues/{issue} + "labeledConversationsCount": "A String", # Number of conversations attached to the issue at this point in time. + }, + }, + "unclassifiedConversationsCount": "A String", # Number of analyzed conversations for which no issue was applicable at this point in time. + }, + "updateTime": "A String", # Output only. The most recent time at which the issue model was updated. + }, + ], +}+++ +patch(name, body=None, updateMask=None, x__xgafv=None)+Updates an issue model. + +Args: + name: string, Immutable. The resource name of the issue model. Format: projects/{project}/locations/{location}/issueModels/{issue_model} (required) + body: object, The request body. + The object takes the form of: + +{ # The issue model resource. + "createTime": "A String", # Output only. The time at which this issue model was created. + "displayName": "A String", # The representative name for the issue model. + "inputDataConfig": { # Configs for the input data used to create the issue model. # Configs for the input data that used to create the issue model. + "medium": "A String", # Required. Medium of conversations used in training data. + "trainingConversationsCount": "A String", # Output only. Number of conversations used in training. Output only. + }, + "name": "A String", # Immutable. The resource name of the issue model. Format: projects/{project}/locations/{location}/issueModels/{issue_model} + "state": "A String", # Output only. State of the model. + "trainingStats": { # Aggregated statistics about an issue model. # Output only. Immutable. The issue model's label statistics on its training data. + "analyzedConversationsCount": "A String", # Number of conversations the issue model has analyzed at this point in time. + "issueStats": { # Statistics on each issue. Key is the issue's resource name. + "a_key": { # Aggregated statistics about an issue. + "issue": "A String", # Issue resource. Format: projects/{project}/locations/{location}/issueModels/{issue_model}/issues/{issue} + "labeledConversationsCount": "A String", # Number of conversations attached to the issue at this point in time. + }, + }, + "unclassifiedConversationsCount": "A String", # Number of analyzed conversations for which no issue was applicable at this point in time. + }, + "updateTime": "A String", # Output only. The most recent time at which the issue model was updated. +} + + updateMask: string, The list of fields to be updated. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The issue model resource. + "createTime": "A String", # Output only. The time at which this issue model was created. + "displayName": "A String", # The representative name for the issue model. + "inputDataConfig": { # Configs for the input data used to create the issue model. # Configs for the input data that used to create the issue model. + "medium": "A String", # Required. Medium of conversations used in training data. + "trainingConversationsCount": "A String", # Output only. Number of conversations used in training. Output only. + }, + "name": "A String", # Immutable. The resource name of the issue model. Format: projects/{project}/locations/{location}/issueModels/{issue_model} + "state": "A String", # Output only. State of the model. + "trainingStats": { # Aggregated statistics about an issue model. # Output only. Immutable. The issue model's label statistics on its training data. + "analyzedConversationsCount": "A String", # Number of conversations the issue model has analyzed at this point in time. + "issueStats": { # Statistics on each issue. Key is the issue's resource name. + "a_key": { # Aggregated statistics about an issue. + "issue": "A String", # Issue resource. Format: projects/{project}/locations/{location}/issueModels/{issue_model}/issues/{issue} + "labeledConversationsCount": "A String", # Number of conversations attached to the issue at this point in time. + }, + }, + "unclassifiedConversationsCount": "A String", # Number of analyzed conversations for which no issue was applicable at this point in time. + }, + "updateTime": "A String", # Output only. The most recent time at which the issue model was updated. +}+++ + \ No newline at end of file diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.issueModels.issues.html b/docs/dyn/contactcenterinsights_v1.projects.locations.issueModels.issues.html new file mode 100644 index 00000000000..c7fb8984b5d --- /dev/null +++ b/docs/dyn/contactcenterinsights_v1.projects.locations.issueModels.issues.html @@ -0,0 +1,176 @@ + + + +undeploy(name, body=None, x__xgafv=None)+Undeploys an issue model. An issue model can not be used in analysis after it has been undeployed. + +Args: + name: string, Required. The issue model to undeploy. (required) + body: object, The request body. + The object takes the form of: + +{ # The request to undeploy an issue model. + "name": "A String", # Required. The issue model to undeploy. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+Contact Center AI Insights API . projects . locations . issueModels . issues
+Instance Methods
++
+close()Close httplib2 connections.
+ +Gets an issue.
+ +Lists issues.
++
+patch(name, body=None, updateMask=None, x__xgafv=None)Updates an issue.
+Method Details
+++ +close()+Close httplib2 connections.+++ +get(name, x__xgafv=None)+Gets an issue. + +Args: + name: string, Required. The name of the issue to get. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The issue resource. + "createTime": "A String", # Output only. The time at which this issue was created. + "displayName": "A String", # The representative name for the issue. + "name": "A String", # Immutable. The resource name of the issue. Format: projects/{project}/locations/{location}/issueModels/{issue_model}/issues/{issue} + "updateTime": "A String", # Output only. The most recent time that this issue was updated. +}+++ +list(parent, x__xgafv=None)+Lists issues. + +Args: + parent: string, Required. The parent resource of the issue. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response of listing issues. + "issues": [ # The issues that match the request. + { # The issue resource. + "createTime": "A String", # Output only. The time at which this issue was created. + "displayName": "A String", # The representative name for the issue. + "name": "A String", # Immutable. The resource name of the issue. Format: projects/{project}/locations/{location}/issueModels/{issue_model}/issues/{issue} + "updateTime": "A String", # Output only. The most recent time that this issue was updated. + }, + ], +}+++ + \ No newline at end of file diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.operations.html b/docs/dyn/contactcenterinsights_v1.projects.locations.operations.html new file mode 100644 index 00000000000..796072f9fa0 --- /dev/null +++ b/docs/dyn/contactcenterinsights_v1.projects.locations.operations.html @@ -0,0 +1,208 @@ + + + +patch(name, body=None, updateMask=None, x__xgafv=None)+Updates an issue. + +Args: + name: string, Immutable. The resource name of the issue. Format: projects/{project}/locations/{location}/issueModels/{issue_model}/issues/{issue} (required) + body: object, The request body. + The object takes the form of: + +{ # The issue resource. + "createTime": "A String", # Output only. The time at which this issue was created. + "displayName": "A String", # The representative name for the issue. + "name": "A String", # Immutable. The resource name of the issue. Format: projects/{project}/locations/{location}/issueModels/{issue_model}/issues/{issue} + "updateTime": "A String", # Output only. The most recent time that this issue was updated. +} + + updateMask: string, The list of fields to be updated. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The issue resource. + "createTime": "A String", # Output only. The time at which this issue was created. + "displayName": "A String", # The representative name for the issue. + "name": "A String", # Immutable. The resource name of the issue. Format: projects/{project}/locations/{location}/issueModels/{issue_model}/issues/{issue} + "updateTime": "A String", # Output only. The most recent time that this issue was updated. +}+Contact Center AI Insights API . projects . locations . operations
+Instance Methods
+ +Starts asynchronous cancellation on a long-running operation. The server makes a best effort to cancel the operation, but success is not guaranteed. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`. Clients can use Operations.GetOperation or other methods to check whether the cancellation succeeded or whether the operation completed despite cancellation. On successful cancellation, the operation is not deleted; instead, it becomes an operation with an Operation.error value with a google.rpc.Status.code of 1, corresponding to `Code.CANCELLED`.
++
+close()Close httplib2 connections.
+ +Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
++
+list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `"/v1/{name=users/*}/operations"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.
++
+list_next(previous_request, previous_response)Retrieves the next page of results.
+Method Details
+++ +cancel(name, x__xgafv=None)+Starts asynchronous cancellation on a long-running operation. The server makes a best effort to cancel the operation, but success is not guaranteed. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`. Clients can use Operations.GetOperation or other methods to check whether the cancellation succeeded or whether the operation completed despite cancellation. On successful cancellation, the operation is not deleted; instead, it becomes an operation with an Operation.error value with a google.rpc.Status.code of 1, corresponding to `Code.CANCELLED`. + +Args: + name: string, The name of the operation resource to be cancelled. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`. +}+++ +close()+Close httplib2 connections.+++ +get(name, x__xgafv=None)+Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service. + +Args: + name: string, The name of the operation resource. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)+Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `"/v1/{name=users/*}/operations"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id. + +Args: + name: string, The name of the operation's parent resource. (required) + filter: string, The standard list filter. + pageSize: integer, The standard list page size. + pageToken: string, The standard list page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response message for Operations.ListOperations. + "nextPageToken": "A String", # The standard List next-page token. + "operations": [ # A list of operations that matches the specified filter in the request. + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + }, + ], +}+++ + \ No newline at end of file diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.phraseMatchers.html b/docs/dyn/contactcenterinsights_v1.projects.locations.phraseMatchers.html new file mode 100644 index 00000000000..8092bb59253 --- /dev/null +++ b/docs/dyn/contactcenterinsights_v1.projects.locations.phraseMatchers.html @@ -0,0 +1,301 @@ + + + +list_next(previous_request, previous_response)+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Contact Center AI Insights API . projects . locations . phraseMatchers
+Instance Methods
++
+close()Close httplib2 connections.
++
+create(parent, body=None, x__xgafv=None)Creates a phrase matcher.
+ +Deletes a phrase matcher.
+ +Gets a phrase matcher.
++
+list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)Lists phrase matchers.
++
+list_next(previous_request, previous_response)Retrieves the next page of results.
+Method Details
+++ +close()+Close httplib2 connections.+++ +create(parent, body=None, x__xgafv=None)+Creates a phrase matcher. + +Args: + parent: string, Required. The parent resource of the phrase matcher. Required. The location to create a phrase matcher for. Format: `projects//locations/` or `projects//locations/` (required) + body: object, The request body. + The object takes the form of: + +{ # The phrase matcher resource. + "activationUpdateTime": "A String", # Output only. The most recent time at which the activation status was updated. + "active": True or False, # Applies the phrase matcher only when it is active. + "displayName": "A String", # The human-readable name of the phrase matcher. + "name": "A String", # The resource name of the phrase matcher. Format: projects/{project}/locations/{location}/phraseMatchers/{phrase_matcher} + "phraseMatchRuleGroups": [ # A list of phase match rule groups that are included in this matcher. + { # A message representing a rule in the phrase matcher. + "phraseMatchRules": [ # A list of phase match rules that are included in this group. + { # The data for a phrase match rule. + "config": { # Configuration information of a phrase match rule. # Provides additional information about the rule that specifies how to apply the rule. + "exactMatchConfig": { # Exact match configuration. # The configuration for the exact match rule. + "caseSensitive": True or False, # Whether to consider case sensitivity when performing an exact match. + }, + }, + "negated": True or False, # Specifies whether the phrase must be missing from the transcript segment or present in the transcript segment. + "query": "A String", # Required. The phrase to be matched. + }, + ], + "type": "A String", # Required. The type of this phrase match rule group. + }, + ], + "revisionCreateTime": "A String", # Output only. The timestamp of when the revision was created. It is also the create time when a new matcher is added. + "revisionId": "A String", # Output only. Immutable. The revision ID of the phrase matcher. A new revision is committed whenever the matcher is changed, except when it is activated or deactivated. A server generated random ID will be used. Example: locations/global/phraseMatchers/my-first-matcher@1234567 + "roleMatch": "A String", # The role whose utterances the phrase matcher should be matched against. If the role is ROLE_UNSPECIFIED it will be matched against any utterances in the transcript. + "type": "A String", # Required. The type of this phrase matcher. + "versionTag": "A String", # The customized version tag to use for the phrase matcher. If not specified, it will default to `revision_id`. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The phrase matcher resource. + "activationUpdateTime": "A String", # Output only. The most recent time at which the activation status was updated. + "active": True or False, # Applies the phrase matcher only when it is active. + "displayName": "A String", # The human-readable name of the phrase matcher. + "name": "A String", # The resource name of the phrase matcher. Format: projects/{project}/locations/{location}/phraseMatchers/{phrase_matcher} + "phraseMatchRuleGroups": [ # A list of phase match rule groups that are included in this matcher. + { # A message representing a rule in the phrase matcher. + "phraseMatchRules": [ # A list of phase match rules that are included in this group. + { # The data for a phrase match rule. + "config": { # Configuration information of a phrase match rule. # Provides additional information about the rule that specifies how to apply the rule. + "exactMatchConfig": { # Exact match configuration. # The configuration for the exact match rule. + "caseSensitive": True or False, # Whether to consider case sensitivity when performing an exact match. + }, + }, + "negated": True or False, # Specifies whether the phrase must be missing from the transcript segment or present in the transcript segment. + "query": "A String", # Required. The phrase to be matched. + }, + ], + "type": "A String", # Required. The type of this phrase match rule group. + }, + ], + "revisionCreateTime": "A String", # Output only. The timestamp of when the revision was created. It is also the create time when a new matcher is added. + "revisionId": "A String", # Output only. Immutable. The revision ID of the phrase matcher. A new revision is committed whenever the matcher is changed, except when it is activated or deactivated. A server generated random ID will be used. Example: locations/global/phraseMatchers/my-first-matcher@1234567 + "roleMatch": "A String", # The role whose utterances the phrase matcher should be matched against. If the role is ROLE_UNSPECIFIED it will be matched against any utterances in the transcript. + "type": "A String", # Required. The type of this phrase matcher. + "versionTag": "A String", # The customized version tag to use for the phrase matcher. If not specified, it will default to `revision_id`. +}+++ +delete(name, x__xgafv=None)+Deletes a phrase matcher. + +Args: + name: string, Required. The name of the phrase matcher to delete. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`. +}+++ +get(name, x__xgafv=None)+Gets a phrase matcher. + +Args: + name: string, Required. The name of the phrase matcher to get. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The phrase matcher resource. + "activationUpdateTime": "A String", # Output only. The most recent time at which the activation status was updated. + "active": True or False, # Applies the phrase matcher only when it is active. + "displayName": "A String", # The human-readable name of the phrase matcher. + "name": "A String", # The resource name of the phrase matcher. Format: projects/{project}/locations/{location}/phraseMatchers/{phrase_matcher} + "phraseMatchRuleGroups": [ # A list of phase match rule groups that are included in this matcher. + { # A message representing a rule in the phrase matcher. + "phraseMatchRules": [ # A list of phase match rules that are included in this group. + { # The data for a phrase match rule. + "config": { # Configuration information of a phrase match rule. # Provides additional information about the rule that specifies how to apply the rule. + "exactMatchConfig": { # Exact match configuration. # The configuration for the exact match rule. + "caseSensitive": True or False, # Whether to consider case sensitivity when performing an exact match. + }, + }, + "negated": True or False, # Specifies whether the phrase must be missing from the transcript segment or present in the transcript segment. + "query": "A String", # Required. The phrase to be matched. + }, + ], + "type": "A String", # Required. The type of this phrase match rule group. + }, + ], + "revisionCreateTime": "A String", # Output only. The timestamp of when the revision was created. It is also the create time when a new matcher is added. + "revisionId": "A String", # Output only. Immutable. The revision ID of the phrase matcher. A new revision is committed whenever the matcher is changed, except when it is activated or deactivated. A server generated random ID will be used. Example: locations/global/phraseMatchers/my-first-matcher@1234567 + "roleMatch": "A String", # The role whose utterances the phrase matcher should be matched against. If the role is ROLE_UNSPECIFIED it will be matched against any utterances in the transcript. + "type": "A String", # Required. The type of this phrase matcher. + "versionTag": "A String", # The customized version tag to use for the phrase matcher. If not specified, it will default to `revision_id`. +}+++ +list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)+Lists phrase matchers. + +Args: + parent: string, Required. The parent resource of the phrase matcher. (required) + filter: string, A filter to reduce results to a specific subset. Useful for querying phrase matchers with specific properties. + pageSize: integer, The maximum number of phrase matchers to return in the response. If this value is zero, the service will select a default size. A call might return fewer objects than requested. A non-empty `next_page_token` in the response indicates that more data is available. + pageToken: string, The value returned by the last `ListPhraseMatchersResponse`. This value indicates that this is a continuation of a prior `ListPhraseMatchers` call and that the system should return the next page of data. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response of listing phrase matchers. + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. + "phraseMatchers": [ # The phrase matchers that match the request. + { # The phrase matcher resource. + "activationUpdateTime": "A String", # Output only. The most recent time at which the activation status was updated. + "active": True or False, # Applies the phrase matcher only when it is active. + "displayName": "A String", # The human-readable name of the phrase matcher. + "name": "A String", # The resource name of the phrase matcher. Format: projects/{project}/locations/{location}/phraseMatchers/{phrase_matcher} + "phraseMatchRuleGroups": [ # A list of phase match rule groups that are included in this matcher. + { # A message representing a rule in the phrase matcher. + "phraseMatchRules": [ # A list of phase match rules that are included in this group. + { # The data for a phrase match rule. + "config": { # Configuration information of a phrase match rule. # Provides additional information about the rule that specifies how to apply the rule. + "exactMatchConfig": { # Exact match configuration. # The configuration for the exact match rule. + "caseSensitive": True or False, # Whether to consider case sensitivity when performing an exact match. + }, + }, + "negated": True or False, # Specifies whether the phrase must be missing from the transcript segment or present in the transcript segment. + "query": "A String", # Required. The phrase to be matched. + }, + ], + "type": "A String", # Required. The type of this phrase match rule group. + }, + ], + "revisionCreateTime": "A String", # Output only. The timestamp of when the revision was created. It is also the create time when a new matcher is added. + "revisionId": "A String", # Output only. Immutable. The revision ID of the phrase matcher. A new revision is committed whenever the matcher is changed, except when it is activated or deactivated. A server generated random ID will be used. Example: locations/global/phraseMatchers/my-first-matcher@1234567 + "roleMatch": "A String", # The role whose utterances the phrase matcher should be matched against. If the role is ROLE_UNSPECIFIED it will be matched against any utterances in the transcript. + "type": "A String", # Required. The type of this phrase matcher. + "versionTag": "A String", # The customized version tag to use for the phrase matcher. If not specified, it will default to `revision_id`. + }, + ], +}+++ + \ No newline at end of file diff --git a/docs/dyn/container_v1.projects.locations.clusters.html b/docs/dyn/container_v1.projects.locations.clusters.html index c7fb79ea9b0..38099515263 100644 --- a/docs/dyn/container_v1.projects.locations.clusters.html +++ b/docs/dyn/container_v1.projects.locations.clusters.html @@ -300,6 +300,7 @@list_next(previous_request, previous_response)+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Method Details
"maxUnavailable": 42, # The maximum number of nodes that can be simultaneously unavailable during the upgrade process. A node is considered available if its status is Ready. }, }, + "autoscalingProfile": "A String", # Defines autoscaling behaviour. "enableNodeAutoprovisioning": True or False, # Enables automatic node pool creation and deletion. "resourceLimits": [ # Contains global constraints regarding minimum and maximum amount of resources in the cluster. { # Contains information about amount of some resource in the cluster. For memory, value should be in GB. @@ -368,6 +369,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -410,6 +418,13 @@Method Details
], "enabled": True or False, # Whether or not master authorized networks is enabled. }, + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. @@ -591,6 +606,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. @@ -870,6 +890,7 @@Method Details
"maxUnavailable": 42, # The maximum number of nodes that can be simultaneously unavailable during the upgrade process. A node is considered available if its status is Ready. }, }, + "autoscalingProfile": "A String", # Defines autoscaling behaviour. "enableNodeAutoprovisioning": True or False, # Enables automatic node pool creation and deletion. "resourceLimits": [ # Contains global constraints regarding minimum and maximum amount of resources in the cluster. { # Contains information about amount of some resource in the cluster. For memory, value should be in GB. @@ -938,6 +959,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -980,6 +1008,13 @@Method Details
], "enabled": True or False, # Whether or not master authorized networks is enabled. }, + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. @@ -1161,6 +1196,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. @@ -1343,6 +1383,7 @@Method Details
"maxUnavailable": 42, # The maximum number of nodes that can be simultaneously unavailable during the upgrade process. A node is considered available if its status is Ready. }, }, + "autoscalingProfile": "A String", # Defines autoscaling behaviour. "enableNodeAutoprovisioning": True or False, # Enables automatic node pool creation and deletion. "resourceLimits": [ # Contains global constraints regarding minimum and maximum amount of resources in the cluster. { # Contains information about amount of some resource in the cluster. For memory, value should be in GB. @@ -1411,6 +1452,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -1453,6 +1501,13 @@Method Details
], "enabled": True or False, # Whether or not master authorized networks is enabled. }, + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. @@ -1634,6 +1689,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. @@ -2628,6 +2688,7 @@Method Details
"maxUnavailable": 42, # The maximum number of nodes that can be simultaneously unavailable during the upgrade process. A node is considered available if its status is Ready. }, }, + "autoscalingProfile": "A String", # Defines autoscaling behaviour. "enableNodeAutoprovisioning": True or False, # Enables automatic node pool creation and deletion. "resourceLimits": [ # Contains global constraints regarding minimum and maximum amount of resources in the cluster. { # Contains information about amount of some resource in the cluster. For memory, value should be in GB. @@ -2655,6 +2716,13 @@Method Details
"desiredLocations": [ # The desired list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This list must always include the cluster's primary zone. Warning: changing cluster locations will update the locations of all node pools and will result in nodes being added and/or removed. "A String", ], + "desiredLoggingConfig": { # LoggingConfig is cluster logging configuration. # The desired logging configuration. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "desiredLoggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "desiredMasterAuthorizedNetworksConfig": { # Configuration options for the master authorized networks feature. Enabled master authorized networks will disallow all external traffic to access Kubernetes master through HTTPS except traffic from the given CIDR blocks, Google Compute Engine Public IPs and Google Prod IPs. # The desired configuration options for master authorized networks feature. "cidrBlocks": [ # cidr_blocks define up to 50 external networks that could access Kubernetes master through HTTPS. @@ -2666,6 +2734,13 @@Method Details
"enabled": True or False, # Whether or not master authorized networks is enabled. }, "desiredMasterVersion": "A String", # The Kubernetes version to change the master to. Users may specify either explicit versions offered by Kubernetes Engine or version aliases, which have the following behavior: - "latest": picks the highest valid Kubernetes version - "1.X": picks the highest valid patch+gke.N patch in the 1.X version - "1.X.Y": picks the highest valid gke.N patch in the 1.X.Y version - "1.X.Y-gke.N": picks an explicit Kubernetes version - "-": picks the default Kubernetes version + "desiredMonitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # The desired monitoring configuration. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "desiredMonitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "desiredNodePoolAutoscaling": { # NodePoolAutoscaling contains information required by cluster autoscaler to adjust the size of the node pool to the current cluster usage. # Autoscaler configuration for the node pool specified in desired_node_pool_id. If there is only one pool in the cluster and desired_node_pool_id is not provided then the change applies to that single node pool. "autoprovisioned": True or False, # Can this node pool be deleted automatically. diff --git a/docs/dyn/container_v1.projects.locations.clusters.nodePools.html b/docs/dyn/container_v1.projects.locations.clusters.nodePools.html index c58b22d4279..8f137ccc9f2 100644 --- a/docs/dyn/container_v1.projects.locations.clusters.nodePools.html +++ b/docs/dyn/container_v1.projects.locations.clusters.nodePools.html @@ -218,6 +218,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. @@ -480,6 +485,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. @@ -608,6 +618,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. diff --git a/docs/dyn/container_v1.projects.zones.clusters.html b/docs/dyn/container_v1.projects.zones.clusters.html index 3305b0a89b4..f0ef7f6c2cd 100644 --- a/docs/dyn/container_v1.projects.zones.clusters.html +++ b/docs/dyn/container_v1.projects.zones.clusters.html @@ -401,6 +401,7 @@Method Details
"maxUnavailable": 42, # The maximum number of nodes that can be simultaneously unavailable during the upgrade process. A node is considered available if its status is Ready. }, }, + "autoscalingProfile": "A String", # Defines autoscaling behaviour. "enableNodeAutoprovisioning": True or False, # Enables automatic node pool creation and deletion. "resourceLimits": [ # Contains global constraints regarding minimum and maximum amount of resources in the cluster. { # Contains information about amount of some resource in the cluster. For memory, value should be in GB. @@ -469,6 +470,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -511,6 +519,13 @@Method Details
], "enabled": True or False, # Whether or not master authorized networks is enabled. }, + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. @@ -692,6 +707,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. @@ -971,6 +991,7 @@Method Details
"maxUnavailable": 42, # The maximum number of nodes that can be simultaneously unavailable during the upgrade process. A node is considered available if its status is Ready. }, }, + "autoscalingProfile": "A String", # Defines autoscaling behaviour. "enableNodeAutoprovisioning": True or False, # Enables automatic node pool creation and deletion. "resourceLimits": [ # Contains global constraints regarding minimum and maximum amount of resources in the cluster. { # Contains information about amount of some resource in the cluster. For memory, value should be in GB. @@ -1039,6 +1060,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -1081,6 +1109,13 @@Method Details
], "enabled": True or False, # Whether or not master authorized networks is enabled. }, + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. @@ -1262,6 +1297,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. @@ -1488,6 +1528,7 @@Method Details
"maxUnavailable": 42, # The maximum number of nodes that can be simultaneously unavailable during the upgrade process. A node is considered available if its status is Ready. }, }, + "autoscalingProfile": "A String", # Defines autoscaling behaviour. "enableNodeAutoprovisioning": True or False, # Enables automatic node pool creation and deletion. "resourceLimits": [ # Contains global constraints regarding minimum and maximum amount of resources in the cluster. { # Contains information about amount of some resource in the cluster. For memory, value should be in GB. @@ -1556,6 +1597,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -1598,6 +1646,13 @@Method Details
], "enabled": True or False, # Whether or not master authorized networks is enabled. }, + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. @@ -1779,6 +1834,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. @@ -2689,6 +2749,7 @@Method Details
"maxUnavailable": 42, # The maximum number of nodes that can be simultaneously unavailable during the upgrade process. A node is considered available if its status is Ready. }, }, + "autoscalingProfile": "A String", # Defines autoscaling behaviour. "enableNodeAutoprovisioning": True or False, # Enables automatic node pool creation and deletion. "resourceLimits": [ # Contains global constraints regarding minimum and maximum amount of resources in the cluster. { # Contains information about amount of some resource in the cluster. For memory, value should be in GB. @@ -2716,6 +2777,13 @@Method Details
"desiredLocations": [ # The desired list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This list must always include the cluster's primary zone. Warning: changing cluster locations will update the locations of all node pools and will result in nodes being added and/or removed. "A String", ], + "desiredLoggingConfig": { # LoggingConfig is cluster logging configuration. # The desired logging configuration. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "desiredLoggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "desiredMasterAuthorizedNetworksConfig": { # Configuration options for the master authorized networks feature. Enabled master authorized networks will disallow all external traffic to access Kubernetes master through HTTPS except traffic from the given CIDR blocks, Google Compute Engine Public IPs and Google Prod IPs. # The desired configuration options for master authorized networks feature. "cidrBlocks": [ # cidr_blocks define up to 50 external networks that could access Kubernetes master through HTTPS. @@ -2727,6 +2795,13 @@Method Details
"enabled": True or False, # Whether or not master authorized networks is enabled. }, "desiredMasterVersion": "A String", # The Kubernetes version to change the master to. Users may specify either explicit versions offered by Kubernetes Engine or version aliases, which have the following behavior: - "latest": picks the highest valid Kubernetes version - "1.X": picks the highest valid patch+gke.N patch in the 1.X version - "1.X.Y": picks the highest valid gke.N patch in the 1.X.Y version - "1.X.Y-gke.N": picks an explicit Kubernetes version - "-": picks the default Kubernetes version + "desiredMonitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # The desired monitoring configuration. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "desiredMonitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "desiredNodePoolAutoscaling": { # NodePoolAutoscaling contains information required by cluster autoscaler to adjust the size of the node pool to the current cluster usage. # Autoscaler configuration for the node pool specified in desired_node_pool_id. If there is only one pool in the cluster and desired_node_pool_id is not provided then the change applies to that single node pool. "autoprovisioned": True or False, # Can this node pool be deleted automatically. diff --git a/docs/dyn/container_v1.projects.zones.clusters.nodePools.html b/docs/dyn/container_v1.projects.zones.clusters.nodePools.html index 7d156e5ff9a..47ba79b2dba 100644 --- a/docs/dyn/container_v1.projects.zones.clusters.nodePools.html +++ b/docs/dyn/container_v1.projects.zones.clusters.nodePools.html @@ -307,6 +307,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. @@ -569,6 +574,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. @@ -697,6 +707,11 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. "status": "A String", # [Output only] The status of the nodes in this pool instance. diff --git a/docs/dyn/container_v1beta1.projects.locations.clusters.html b/docs/dyn/container_v1beta1.projects.locations.clusters.html index a47fa6c5209..d1d0d419e62 100644 --- a/docs/dyn/container_v1beta1.projects.locations.clusters.html +++ b/docs/dyn/container_v1beta1.projects.locations.clusters.html @@ -351,6 +351,9 @@Method Details
"endpoint": "A String", # [Output only] The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "expireTime": "A String", # [Output only] The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. "id": "A String", # Output only. Unique id for the cluster. + "identityServiceConfig": { # IdentityServiceConfig is configuration for Identity Service which allows customers to use external identity providers with the K8S API # Configuration for Identity Service component. + "enabled": True or False, # Whether to enable the Identity Service component + }, "initialClusterVersion": "A String", # The initial Kubernetes version for this cluster. Valid versions are those found in validMasterVersions returned by getServerConfig. The version can be upgraded over time; such upgrades are reflected in currentMasterVersion and currentNodeVersion. Users may specify either explicit versions offered by Kubernetes Engine or version aliases, which have the following behavior: - "latest": picks the highest valid Kubernetes version - "1.X": picks the highest valid patch+gke.N patch in the 1.X version - "1.X.Y": picks the highest valid gke.N patch in the 1.X.Y version - "1.X.Y-gke.N": picks an explicit Kubernetes version - "","-": picks the default Kubernetes version "initialNodeCount": 42, # The number of nodes to create in this cluster. You must ensure that your Compute Engine [resource quota](https://cloud.google.com/compute/quotas) is sufficient for this number of instances. You must also have available firewall and routes quota. For requests, this field should only be used in lieu of a "node_pool" object, since this configuration (along with the "node_config") will be used to create a "NodePool" object with an auto-generated name. Do not use this and a node_pool at the same time. This field is deprecated, use node_pool.initial_node_count instead. "instanceGroupUrls": [ # Deprecated. Use node_pools.instance_group_urls. @@ -380,6 +383,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -425,6 +435,13 @@Method Details
"enabled": True or False, # Whether or not master authorized networks is enabled. }, "masterIpv4CidrBlock": "A String", # The IP prefix in CIDR notation to use for the hosted master network. This prefix will be used for assigning private IP addresses to the master or set of masters, as well as the ILB VIP. This field is deprecated, use private_cluster_config.master_ipv4_cidr_block instead. + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. On output this shows the network ID instead of the name. @@ -630,10 +647,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. @@ -979,6 +996,9 @@Method Details
"endpoint": "A String", # [Output only] The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "expireTime": "A String", # [Output only] The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. "id": "A String", # Output only. Unique id for the cluster. + "identityServiceConfig": { # IdentityServiceConfig is configuration for Identity Service which allows customers to use external identity providers with the K8S API # Configuration for Identity Service component. + "enabled": True or False, # Whether to enable the Identity Service component + }, "initialClusterVersion": "A String", # The initial Kubernetes version for this cluster. Valid versions are those found in validMasterVersions returned by getServerConfig. The version can be upgraded over time; such upgrades are reflected in currentMasterVersion and currentNodeVersion. Users may specify either explicit versions offered by Kubernetes Engine or version aliases, which have the following behavior: - "latest": picks the highest valid Kubernetes version - "1.X": picks the highest valid patch+gke.N patch in the 1.X version - "1.X.Y": picks the highest valid gke.N patch in the 1.X.Y version - "1.X.Y-gke.N": picks an explicit Kubernetes version - "","-": picks the default Kubernetes version "initialNodeCount": 42, # The number of nodes to create in this cluster. You must ensure that your Compute Engine [resource quota](https://cloud.google.com/compute/quotas) is sufficient for this number of instances. You must also have available firewall and routes quota. For requests, this field should only be used in lieu of a "node_pool" object, since this configuration (along with the "node_config") will be used to create a "NodePool" object with an auto-generated name. Do not use this and a node_pool at the same time. This field is deprecated, use node_pool.initial_node_count instead. "instanceGroupUrls": [ # Deprecated. Use node_pools.instance_group_urls. @@ -1008,6 +1028,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -1053,6 +1080,13 @@Method Details
"enabled": True or False, # Whether or not master authorized networks is enabled. }, "masterIpv4CidrBlock": "A String", # The IP prefix in CIDR notation to use for the hosted master network. This prefix will be used for assigning private IP addresses to the master or set of masters, as well as the ILB VIP. This field is deprecated, use private_cluster_config.master_ipv4_cidr_block instead. + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. On output this shows the network ID instead of the name. @@ -1258,10 +1292,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. @@ -1510,6 +1544,9 @@Method Details
"endpoint": "A String", # [Output only] The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "expireTime": "A String", # [Output only] The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. "id": "A String", # Output only. Unique id for the cluster. + "identityServiceConfig": { # IdentityServiceConfig is configuration for Identity Service which allows customers to use external identity providers with the K8S API # Configuration for Identity Service component. + "enabled": True or False, # Whether to enable the Identity Service component + }, "initialClusterVersion": "A String", # The initial Kubernetes version for this cluster. Valid versions are those found in validMasterVersions returned by getServerConfig. The version can be upgraded over time; such upgrades are reflected in currentMasterVersion and currentNodeVersion. Users may specify either explicit versions offered by Kubernetes Engine or version aliases, which have the following behavior: - "latest": picks the highest valid Kubernetes version - "1.X": picks the highest valid patch+gke.N patch in the 1.X version - "1.X.Y": picks the highest valid gke.N patch in the 1.X.Y version - "1.X.Y-gke.N": picks an explicit Kubernetes version - "","-": picks the default Kubernetes version "initialNodeCount": 42, # The number of nodes to create in this cluster. You must ensure that your Compute Engine [resource quota](https://cloud.google.com/compute/quotas) is sufficient for this number of instances. You must also have available firewall and routes quota. For requests, this field should only be used in lieu of a "node_pool" object, since this configuration (along with the "node_config") will be used to create a "NodePool" object with an auto-generated name. Do not use this and a node_pool at the same time. This field is deprecated, use node_pool.initial_node_count instead. "instanceGroupUrls": [ # Deprecated. Use node_pools.instance_group_urls. @@ -1539,6 +1576,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -1584,6 +1628,13 @@Method Details
"enabled": True or False, # Whether or not master authorized networks is enabled. }, "masterIpv4CidrBlock": "A String", # The IP prefix in CIDR notation to use for the hosted master network. This prefix will be used for assigning private IP addresses to the master or set of masters, as well as the ILB VIP. This field is deprecated, use private_cluster_config.master_ipv4_cidr_block instead. + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. On output this shows the network ID instead of the name. @@ -1789,10 +1840,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. @@ -2842,6 +2893,9 @@Method Details
"clusterDnsDomain": "A String", # cluster_dns_domain is the suffix used for all cluster service records. "clusterDnsScope": "A String", # cluster_dns_scope indicates the scope of access to cluster DNS records. }, + "desiredIdentityServiceConfig": { # IdentityServiceConfig is configuration for Identity Service which allows customers to use external identity providers with the K8S API # The desired Identity Service component configuration. + "enabled": True or False, # Whether to enable the Identity Service component + }, "desiredImageType": "A String", # The desired image type for the node pool. NOTE: Set the "desired_node_pool" field as well. "desiredIntraNodeVisibilityConfig": { # IntraNodeVisibilityConfig contains the desired config of the intra-node visibility on this cluster. # The desired config of Intra-node visibility. "enabled": True or False, # Enables intra node visibility for this cluster. @@ -2852,6 +2906,13 @@Method Details
"desiredLocations": [ # The desired list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This list must always include the cluster's primary zone. Warning: changing cluster locations will update the locations of all node pools and will result in nodes being added and/or removed. "A String", ], + "desiredLoggingConfig": { # LoggingConfig is cluster logging configuration. # The desired logging configuration. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "desiredLoggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "desiredMaster": { # Master is the configuration for components on master. # Configuration for master components. }, @@ -2865,6 +2926,13 @@Method Details
"enabled": True or False, # Whether or not master authorized networks is enabled. }, "desiredMasterVersion": "A String", # The Kubernetes version to change the master to. The only valid value is the latest supported version. Users may specify either explicit versions offered by Kubernetes Engine or version aliases, which have the following behavior: - "latest": picks the highest valid Kubernetes version - "1.X": picks the highest valid patch+gke.N patch in the 1.X version - "1.X.Y": picks the highest valid gke.N patch in the 1.X.Y version - "1.X.Y-gke.N": picks an explicit Kubernetes version - "-": picks the default Kubernetes version + "desiredMonitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # The desired monitoring configuration. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "desiredMonitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "desiredNodePoolAutoscaling": { # NodePoolAutoscaling contains information required by cluster autoscaler to adjust the size of the node pool to the current cluster usage. # Autoscaler configuration for the node pool specified in desired_node_pool_id. If there is only one pool in the cluster and desired_node_pool_id is not provided then the change applies to that single node pool. "autoprovisioned": True or False, # Can this node pool be deleted automatically. diff --git a/docs/dyn/container_v1beta1.projects.locations.clusters.nodePools.html b/docs/dyn/container_v1beta1.projects.locations.clusters.nodePools.html index 5d0626f6ed3..e1b6f7695b3 100644 --- a/docs/dyn/container_v1beta1.projects.locations.clusters.nodePools.html +++ b/docs/dyn/container_v1beta1.projects.locations.clusters.nodePools.html @@ -224,10 +224,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. @@ -497,10 +497,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. @@ -636,10 +636,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. diff --git a/docs/dyn/container_v1beta1.projects.zones.clusters.html b/docs/dyn/container_v1beta1.projects.zones.clusters.html index fd24abf3069..316988269b5 100644 --- a/docs/dyn/container_v1beta1.projects.zones.clusters.html +++ b/docs/dyn/container_v1beta1.projects.zones.clusters.html @@ -459,6 +459,9 @@Method Details
"endpoint": "A String", # [Output only] The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "expireTime": "A String", # [Output only] The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. "id": "A String", # Output only. Unique id for the cluster. + "identityServiceConfig": { # IdentityServiceConfig is configuration for Identity Service which allows customers to use external identity providers with the K8S API # Configuration for Identity Service component. + "enabled": True or False, # Whether to enable the Identity Service component + }, "initialClusterVersion": "A String", # The initial Kubernetes version for this cluster. Valid versions are those found in validMasterVersions returned by getServerConfig. The version can be upgraded over time; such upgrades are reflected in currentMasterVersion and currentNodeVersion. Users may specify either explicit versions offered by Kubernetes Engine or version aliases, which have the following behavior: - "latest": picks the highest valid Kubernetes version - "1.X": picks the highest valid patch+gke.N patch in the 1.X version - "1.X.Y": picks the highest valid gke.N patch in the 1.X.Y version - "1.X.Y-gke.N": picks an explicit Kubernetes version - "","-": picks the default Kubernetes version "initialNodeCount": 42, # The number of nodes to create in this cluster. You must ensure that your Compute Engine [resource quota](https://cloud.google.com/compute/quotas) is sufficient for this number of instances. You must also have available firewall and routes quota. For requests, this field should only be used in lieu of a "node_pool" object, since this configuration (along with the "node_config") will be used to create a "NodePool" object with an auto-generated name. Do not use this and a node_pool at the same time. This field is deprecated, use node_pool.initial_node_count instead. "instanceGroupUrls": [ # Deprecated. Use node_pools.instance_group_urls. @@ -488,6 +491,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -533,6 +543,13 @@Method Details
"enabled": True or False, # Whether or not master authorized networks is enabled. }, "masterIpv4CidrBlock": "A String", # The IP prefix in CIDR notation to use for the hosted master network. This prefix will be used for assigning private IP addresses to the master or set of masters, as well as the ILB VIP. This field is deprecated, use private_cluster_config.master_ipv4_cidr_block instead. + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. On output this shows the network ID instead of the name. @@ -738,10 +755,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. @@ -1087,6 +1104,9 @@Method Details
"endpoint": "A String", # [Output only] The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "expireTime": "A String", # [Output only] The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. "id": "A String", # Output only. Unique id for the cluster. + "identityServiceConfig": { # IdentityServiceConfig is configuration for Identity Service which allows customers to use external identity providers with the K8S API # Configuration for Identity Service component. + "enabled": True or False, # Whether to enable the Identity Service component + }, "initialClusterVersion": "A String", # The initial Kubernetes version for this cluster. Valid versions are those found in validMasterVersions returned by getServerConfig. The version can be upgraded over time; such upgrades are reflected in currentMasterVersion and currentNodeVersion. Users may specify either explicit versions offered by Kubernetes Engine or version aliases, which have the following behavior: - "latest": picks the highest valid Kubernetes version - "1.X": picks the highest valid patch+gke.N patch in the 1.X version - "1.X.Y": picks the highest valid gke.N patch in the 1.X.Y version - "1.X.Y-gke.N": picks an explicit Kubernetes version - "","-": picks the default Kubernetes version "initialNodeCount": 42, # The number of nodes to create in this cluster. You must ensure that your Compute Engine [resource quota](https://cloud.google.com/compute/quotas) is sufficient for this number of instances. You must also have available firewall and routes quota. For requests, this field should only be used in lieu of a "node_pool" object, since this configuration (along with the "node_config") will be used to create a "NodePool" object with an auto-generated name. Do not use this and a node_pool at the same time. This field is deprecated, use node_pool.initial_node_count instead. "instanceGroupUrls": [ # Deprecated. Use node_pools.instance_group_urls. @@ -1116,6 +1136,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -1161,6 +1188,13 @@Method Details
"enabled": True or False, # Whether or not master authorized networks is enabled. }, "masterIpv4CidrBlock": "A String", # The IP prefix in CIDR notation to use for the hosted master network. This prefix will be used for assigning private IP addresses to the master or set of masters, as well as the ILB VIP. This field is deprecated, use private_cluster_config.master_ipv4_cidr_block instead. + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. On output this shows the network ID instead of the name. @@ -1366,10 +1400,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. @@ -1662,6 +1696,9 @@Method Details
"endpoint": "A String", # [Output only] The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "expireTime": "A String", # [Output only] The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. "id": "A String", # Output only. Unique id for the cluster. + "identityServiceConfig": { # IdentityServiceConfig is configuration for Identity Service which allows customers to use external identity providers with the K8S API # Configuration for Identity Service component. + "enabled": True or False, # Whether to enable the Identity Service component + }, "initialClusterVersion": "A String", # The initial Kubernetes version for this cluster. Valid versions are those found in validMasterVersions returned by getServerConfig. The version can be upgraded over time; such upgrades are reflected in currentMasterVersion and currentNodeVersion. Users may specify either explicit versions offered by Kubernetes Engine or version aliases, which have the following behavior: - "latest": picks the highest valid Kubernetes version - "1.X": picks the highest valid patch+gke.N patch in the 1.X version - "1.X.Y": picks the highest valid gke.N patch in the 1.X.Y version - "1.X.Y-gke.N": picks an explicit Kubernetes version - "","-": picks the default Kubernetes version "initialNodeCount": 42, # The number of nodes to create in this cluster. You must ensure that your Compute Engine [resource quota](https://cloud.google.com/compute/quotas) is sufficient for this number of instances. You must also have available firewall and routes quota. For requests, this field should only be used in lieu of a "node_pool" object, since this configuration (along with the "node_config") will be used to create a "NodePool" object with an auto-generated name. Do not use this and a node_pool at the same time. This field is deprecated, use node_pool.initial_node_count instead. "instanceGroupUrls": [ # Deprecated. Use node_pools.instance_group_urls. @@ -1691,6 +1728,13 @@Method Details
"locations": [ # The list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This field provides a default value if [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) are not specified during node pool creation. Warning: changing cluster locations will update the [NodePool.Locations](https://cloud.google.com/kubernetes-engine/docs/reference/rest/v1/projects.locations.clusters.nodePools#NodePool.FIELDS.locations) of all node pools and will result in nodes being added and/or removed. "A String", ], + "loggingConfig": { # LoggingConfig is cluster logging configuration. # Logging configuration for the cluster. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "loggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "maintenancePolicy": { # MaintenancePolicy defines the maintenance policy to be used for the cluster. # Configure the maintenance policy for this cluster. "resourceVersion": "A String", # A hash identifying the version of this policy, so that updates to fields of the policy won't accidentally undo intermediate changes (and so that users of the API unaware of some fields won't accidentally remove other fields). Make a `get()` request to the cluster to get the current resource version and include it with requests to set the policy. @@ -1736,6 +1780,13 @@Method Details
"enabled": True or False, # Whether or not master authorized networks is enabled. }, "masterIpv4CidrBlock": "A String", # The IP prefix in CIDR notation to use for the hosted master network. This prefix will be used for assigning private IP addresses to the master or set of masters, as well as the ILB VIP. This field is deprecated, use private_cluster_config.master_ipv4_cidr_block instead. + "monitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # Monitoring configuration for the cluster. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. "network": "A String", # The name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the cluster is connected. If left unspecified, the `default` network will be used. On output this shows the network ID instead of the name. @@ -1941,10 +1992,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. @@ -2903,6 +2954,9 @@Method Details
"clusterDnsDomain": "A String", # cluster_dns_domain is the suffix used for all cluster service records. "clusterDnsScope": "A String", # cluster_dns_scope indicates the scope of access to cluster DNS records. }, + "desiredIdentityServiceConfig": { # IdentityServiceConfig is configuration for Identity Service which allows customers to use external identity providers with the K8S API # The desired Identity Service component configuration. + "enabled": True or False, # Whether to enable the Identity Service component + }, "desiredImageType": "A String", # The desired image type for the node pool. NOTE: Set the "desired_node_pool" field as well. "desiredIntraNodeVisibilityConfig": { # IntraNodeVisibilityConfig contains the desired config of the intra-node visibility on this cluster. # The desired config of Intra-node visibility. "enabled": True or False, # Enables intra node visibility for this cluster. @@ -2913,6 +2967,13 @@Method Details
"desiredLocations": [ # The desired list of Google Compute Engine [zones](https://cloud.google.com/compute/docs/zones#available) in which the cluster's nodes should be located. This list must always include the cluster's primary zone. Warning: changing cluster locations will update the locations of all node pools and will result in nodes being added and/or removed. "A String", ], + "desiredLoggingConfig": { # LoggingConfig is cluster logging configuration. # The desired logging configuration. + "componentConfig": { # LoggingComponentConfig is cluster logging component configuration. # Logging components configuration + "enableComponents": [ # Select components to collect logs. An empty set would disable all logging. + "A String", + ], + }, + }, "desiredLoggingService": "A String", # The logging service the cluster should use to write logs. Currently available options: * `logging.googleapis.com/kubernetes` - The Cloud Logging service with a Kubernetes-native resource model * `logging.googleapis.com` - The legacy Cloud Logging service (no longer available as of GKE 1.15). * `none` - no logs will be exported from the cluster. If left as an empty string,`logging.googleapis.com/kubernetes` will be used for GKE 1.14+ or `logging.googleapis.com` for earlier versions. "desiredMaster": { # Master is the configuration for components on master. # Configuration for master components. }, @@ -2926,6 +2987,13 @@Method Details
"enabled": True or False, # Whether or not master authorized networks is enabled. }, "desiredMasterVersion": "A String", # The Kubernetes version to change the master to. The only valid value is the latest supported version. Users may specify either explicit versions offered by Kubernetes Engine or version aliases, which have the following behavior: - "latest": picks the highest valid Kubernetes version - "1.X": picks the highest valid patch+gke.N patch in the 1.X version - "1.X.Y": picks the highest valid gke.N patch in the 1.X.Y version - "1.X.Y-gke.N": picks an explicit Kubernetes version - "-": picks the default Kubernetes version + "desiredMonitoringConfig": { # MonitoringConfig is cluster monitoring configuration. # The desired monitoring configuration. + "componentConfig": { # MonitoringComponentConfig is cluster monitoring component configuration. # Monitoring components configuration + "enableComponents": [ # Select components to collect metrics. An empty set would disable all monitoring. + "A String", + ], + }, + }, "desiredMonitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "desiredNodePoolAutoscaling": { # NodePoolAutoscaling contains information required by cluster autoscaler to adjust the size of the node pool to the current cluster usage. # Autoscaler configuration for the node pool specified in desired_node_pool_id. If there is only one pool in the cluster and desired_node_pool_id is not provided then the change applies to that single node pool. "autoprovisioned": True or False, # Can this node pool be deleted automatically. diff --git a/docs/dyn/container_v1beta1.projects.zones.clusters.nodePools.html b/docs/dyn/container_v1beta1.projects.zones.clusters.nodePools.html index f26606677db..8531dfd1542 100644 --- a/docs/dyn/container_v1beta1.projects.zones.clusters.nodePools.html +++ b/docs/dyn/container_v1beta1.projects.zones.clusters.nodePools.html @@ -313,10 +313,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. @@ -586,10 +586,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. @@ -725,10 +725,10 @@Method Details
"maxPodsPerNode": "A String", # Constraint enforced on the max num of pods per node. }, "name": "A String", # The name of the node pool. - "networkConfig": { # Parameters for node pool-level network config. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. - "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. - "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. - "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. + "networkConfig": { # Parameters for node pool-level network config. # Networking configuration for this NodePool. If specified, it overrides the cluster-level defaults. + "createPodRange": True or False, # Input only. Whether to create a new range for pod IPs in this node pool. Defaults are provided for `pod_range` and `pod_ipv4_cidr_block` if they are not specified. If neither `create_pod_range` or `pod_range` are specified, the cluster-level default (`ip_allocation_policy.cluster_ipv4_cidr_block`) is used. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. + "podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created. }, "podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool. "selfLink": "A String", # [Output only] Server-defined URL for the resource. diff --git a/docs/dyn/containeranalysis_v1alpha1.projects.notes.html b/docs/dyn/containeranalysis_v1alpha1.projects.notes.html index 01dfe5cb907..177c2c43341 100644 --- a/docs/dyn/containeranalysis_v1alpha1.projects.notes.html +++ b/docs/dyn/containeranalysis_v1alpha1.projects.notes.html @@ -175,6 +175,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -204,7 +209,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. @@ -330,6 +375,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -359,7 +409,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. @@ -508,6 +598,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -537,7 +632,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. @@ -718,6 +853,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -747,7 +887,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. @@ -889,6 +1069,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -918,7 +1103,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. @@ -1043,6 +1268,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -1072,7 +1302,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. diff --git a/docs/dyn/containeranalysis_v1alpha1.projects.notes.occurrences.html b/docs/dyn/containeranalysis_v1alpha1.projects.notes.occurrences.html index e172ba18d51..46e76405e07 100644 --- a/docs/dyn/containeranalysis_v1alpha1.projects.notes.occurrences.html +++ b/docs/dyn/containeranalysis_v1alpha1.projects.notes.occurrences.html @@ -118,6 +118,36 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. + "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -327,6 +357,70 @@Method Details
}, }, }, + "dsseAttestation": { # An occurrence describing an attestation on a resource # This represents a DSSE attestation occurrence + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # If doing something security critical, make sure to verify the signatures in this metadata. + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, + "statement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". + "predicateType": "A String", # "https://in-toto.io/Provenance/v0.1" for InTotoProvenance. + "provenance": { + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ + { + "digest": { # "": "" + "a_key": "A String", + }, + "name": "A String", + }, + ], + "type": "A String", # Always "https://in-toto.io/Statement/v0.1". + }, + }, + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # https://github.com/secure-systems-lab/dsse + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, "installation": { # This represents how a particular software package may be installed on a system. # Describes the installation of a package on the linked resource. "location": [ # All of the places within the filesystem versions of this package have been found. { # An occurrence of a particular package installation found within a system's filesystem. e.g. glibc was found in /var/lib/dpkg/status @@ -356,6 +450,52 @@Method Details
"uri": "A String", # The unique URI of the resource. E.g., "https://gcr.io/project/image@sha256:foo" for a Docker image. }, "resourceUrl": "A String", # The unique URL of the image or the container for which the `Occurrence` applies. For example, https://gcr.io/project/image@sha256:foo This field can be used as a filter in list requests. + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific software bill of materials document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific relationship between SPDX elements. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this `Occurrence` was last updated. "upgrade": { # An Upgrade Occurrence represents that a specific resource_url could install a specific upgrade. This presence is supplied via local sources (i.e. it is present in the mirror and the running system has noticed its availability). # Describes an upgrade. "distribution": { # The Upgrade Distribution represents metadata about the Upgrade for each operating system (CPE). Some distributions have additional metadata around updates, classifying them into various categories and severities. # Metadata about the upgrade for available for the specific operating system for the resource_url. This allows efficient filtering, as well as making it easier to use the occurrence. diff --git a/docs/dyn/containeranalysis_v1alpha1.projects.occurrences.html b/docs/dyn/containeranalysis_v1alpha1.projects.occurrences.html index 1cd4cf88105..79bfb44856c 100644 --- a/docs/dyn/containeranalysis_v1alpha1.projects.occurrences.html +++ b/docs/dyn/containeranalysis_v1alpha1.projects.occurrences.html @@ -134,6 +134,36 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. + "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -343,6 +373,70 @@Method Details
}, }, }, + "dsseAttestation": { # An occurrence describing an attestation on a resource # This represents a DSSE attestation occurrence + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # If doing something security critical, make sure to verify the signatures in this metadata. + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, + "statement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". + "predicateType": "A String", # "https://in-toto.io/Provenance/v0.1" for InTotoProvenance. + "provenance": { + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ + { + "digest": { # "": "" + "a_key": "A String", + }, + "name": "A String", + }, + ], + "type": "A String", # Always "https://in-toto.io/Statement/v0.1". + }, + }, + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # https://github.com/secure-systems-lab/dsse + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, "installation": { # This represents how a particular software package may be installed on a system. # Describes the installation of a package on the linked resource. "location": [ # All of the places within the filesystem versions of this package have been found. { # An occurrence of a particular package installation found within a system's filesystem. e.g. glibc was found in /var/lib/dpkg/status @@ -372,6 +466,52 @@Method Details
"uri": "A String", # The unique URI of the resource. E.g., "https://gcr.io/project/image@sha256:foo" for a Docker image. }, "resourceUrl": "A String", # The unique URL of the image or the container for which the `Occurrence` applies. For example, https://gcr.io/project/image@sha256:foo This field can be used as a filter in list requests. + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific software bill of materials document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific relationship between SPDX elements. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this `Occurrence` was last updated. "upgrade": { # An Upgrade Occurrence represents that a specific resource_url could install a specific upgrade. This presence is supplied via local sources (i.e. it is present in the mirror and the running system has noticed its availability). # Describes an upgrade. "distribution": { # The Upgrade Distribution represents metadata about the Upgrade for each operating system (CPE). Some distributions have additional metadata around updates, classifying them into various categories and severities. # Metadata about the upgrade for available for the specific operating system for the resource_url. This allows efficient filtering, as well as making it easier to use the occurrence. @@ -444,6 +584,36 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. + "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -653,6 +823,70 @@Method Details
}, }, }, + "dsseAttestation": { # An occurrence describing an attestation on a resource # This represents a DSSE attestation occurrence + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # If doing something security critical, make sure to verify the signatures in this metadata. + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, + "statement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". + "predicateType": "A String", # "https://in-toto.io/Provenance/v0.1" for InTotoProvenance. + "provenance": { + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ + { + "digest": { # "": "" + "a_key": "A String", + }, + "name": "A String", + }, + ], + "type": "A String", # Always "https://in-toto.io/Statement/v0.1". + }, + }, + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # https://github.com/secure-systems-lab/dsse + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, "installation": { # This represents how a particular software package may be installed on a system. # Describes the installation of a package on the linked resource. "location": [ # All of the places within the filesystem versions of this package have been found. { # An occurrence of a particular package installation found within a system's filesystem. e.g. glibc was found in /var/lib/dpkg/status @@ -682,6 +916,52 @@Method Details
"uri": "A String", # The unique URI of the resource. E.g., "https://gcr.io/project/image@sha256:foo" for a Docker image. }, "resourceUrl": "A String", # The unique URL of the image or the container for which the `Occurrence` applies. For example, https://gcr.io/project/image@sha256:foo This field can be used as a filter in list requests. + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific software bill of materials document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific relationship between SPDX elements. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this `Occurrence` was last updated. "upgrade": { # An Upgrade Occurrence represents that a specific resource_url could install a specific upgrade. This presence is supplied via local sources (i.e. it is present in the mirror and the running system has noticed its availability). # Describes an upgrade. "distribution": { # The Upgrade Distribution represents metadata about the Upgrade for each operating system (CPE). Some distributions have additional metadata around updates, classifying them into various categories and severities. # Metadata about the upgrade for available for the specific operating system for the resource_url. This allows efficient filtering, as well as making it easier to use the occurrence. @@ -778,6 +1058,36 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. + "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -987,6 +1297,70 @@Method Details
}, }, }, + "dsseAttestation": { # An occurrence describing an attestation on a resource # This represents a DSSE attestation occurrence + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # If doing something security critical, make sure to verify the signatures in this metadata. + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, + "statement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". + "predicateType": "A String", # "https://in-toto.io/Provenance/v0.1" for InTotoProvenance. + "provenance": { + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ + { + "digest": { # "": "" + "a_key": "A String", + }, + "name": "A String", + }, + ], + "type": "A String", # Always "https://in-toto.io/Statement/v0.1". + }, + }, + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # https://github.com/secure-systems-lab/dsse + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, "installation": { # This represents how a particular software package may be installed on a system. # Describes the installation of a package on the linked resource. "location": [ # All of the places within the filesystem versions of this package have been found. { # An occurrence of a particular package installation found within a system's filesystem. e.g. glibc was found in /var/lib/dpkg/status @@ -1016,6 +1390,52 @@Method Details
"uri": "A String", # The unique URI of the resource. E.g., "https://gcr.io/project/image@sha256:foo" for a Docker image. }, "resourceUrl": "A String", # The unique URL of the image or the container for which the `Occurrence` applies. For example, https://gcr.io/project/image@sha256:foo This field can be used as a filter in list requests. + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific software bill of materials document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific relationship between SPDX elements. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this `Occurrence` was last updated. "upgrade": { # An Upgrade Occurrence represents that a specific resource_url could install a specific upgrade. This presence is supplied via local sources (i.e. it is present in the mirror and the running system has noticed its availability). # Describes an upgrade. "distribution": { # The Upgrade Distribution represents metadata about the Upgrade for each operating system (CPE). Some distributions have additional metadata around updates, classifying them into various categories and severities. # Metadata about the upgrade for available for the specific operating system for the resource_url. This allows efficient filtering, as well as making it easier to use the occurrence. @@ -1179,6 +1599,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -1208,7 +1633,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. @@ -1318,6 +1783,11 @@Method Details
ATTESTATION_AUTHORITY - This represents a logical "role" that can attest to artifacts. UPGRADE - This represents an available software upgrade. COMPLIANCE - This represents a compliance check that can be applied to a resource. + SBOM - This represents a software bill of materials. + SPDX_PACKAGE - This represents an SPDX Package. + SPDX_FILE - This represents an SPDX File. + SPDX_RELATIONSHIP - This represents an SPDX Relationship. + DSSE_ATTESTATION - This represents a DSSE attestation Note name: string, The name field contains the project Id. For example: "projects/{project_id} @Deprecated pageSize: integer, Number of occurrences to return in the list. pageToken: string, Token to provide to skip to a particular spot in the list. @@ -1341,6 +1811,36 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. + "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -1550,6 +2050,70 @@Method Details
}, }, }, + "dsseAttestation": { # An occurrence describing an attestation on a resource # This represents a DSSE attestation occurrence + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # If doing something security critical, make sure to verify the signatures in this metadata. + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, + "statement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". + "predicateType": "A String", # "https://in-toto.io/Provenance/v0.1" for InTotoProvenance. + "provenance": { + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ + { + "digest": { # "": "" + "a_key": "A String", + }, + "name": "A String", + }, + ], + "type": "A String", # Always "https://in-toto.io/Statement/v0.1". + }, + }, + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # https://github.com/secure-systems-lab/dsse + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, "installation": { # This represents how a particular software package may be installed on a system. # Describes the installation of a package on the linked resource. "location": [ # All of the places within the filesystem versions of this package have been found. { # An occurrence of a particular package installation found within a system's filesystem. e.g. glibc was found in /var/lib/dpkg/status @@ -1579,6 +2143,52 @@Method Details
"uri": "A String", # The unique URI of the resource. E.g., "https://gcr.io/project/image@sha256:foo" for a Docker image. }, "resourceUrl": "A String", # The unique URL of the image or the container for which the `Occurrence` applies. For example, https://gcr.io/project/image@sha256:foo This field can be used as a filter in list requests. + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific software bill of materials document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific relationship between SPDX elements. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this `Occurrence` was last updated. "upgrade": { # An Upgrade Occurrence represents that a specific resource_url could install a specific upgrade. This presence is supplied via local sources (i.e. it is present in the mirror and the running system has noticed its availability). # Describes an upgrade. "distribution": { # The Upgrade Distribution represents metadata about the Upgrade for each operating system (CPE). Some distributions have additional metadata around updates, classifying them into various categories and severities. # Metadata about the upgrade for available for the specific operating system for the resource_url. This allows efficient filtering, as well as making it easier to use the occurrence. @@ -1668,6 +2278,36 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. + "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -1877,6 +2517,70 @@Method Details
}, }, }, + "dsseAttestation": { # An occurrence describing an attestation on a resource # This represents a DSSE attestation occurrence + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # If doing something security critical, make sure to verify the signatures in this metadata. + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, + "statement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". + "predicateType": "A String", # "https://in-toto.io/Provenance/v0.1" for InTotoProvenance. + "provenance": { + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ + { + "digest": { # "": "" + "a_key": "A String", + }, + "name": "A String", + }, + ], + "type": "A String", # Always "https://in-toto.io/Statement/v0.1". + }, + }, + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # https://github.com/secure-systems-lab/dsse + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, "installation": { # This represents how a particular software package may be installed on a system. # Describes the installation of a package on the linked resource. "location": [ # All of the places within the filesystem versions of this package have been found. { # An occurrence of a particular package installation found within a system's filesystem. e.g. glibc was found in /var/lib/dpkg/status @@ -1906,6 +2610,52 @@Method Details
"uri": "A String", # The unique URI of the resource. E.g., "https://gcr.io/project/image@sha256:foo" for a Docker image. }, "resourceUrl": "A String", # The unique URL of the image or the container for which the `Occurrence` applies. For example, https://gcr.io/project/image@sha256:foo This field can be used as a filter in list requests. + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific software bill of materials document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific relationship between SPDX elements. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this `Occurrence` was last updated. "upgrade": { # An Upgrade Occurrence represents that a specific resource_url could install a specific upgrade. This presence is supplied via local sources (i.e. it is present in the mirror and the running system has noticed its availability). # Describes an upgrade. "distribution": { # The Upgrade Distribution represents metadata about the Upgrade for each operating system (CPE). Some distributions have additional metadata around updates, classifying them into various categories and severities. # Metadata about the upgrade for available for the specific operating system for the resource_url. This allows efficient filtering, as well as making it easier to use the occurrence. @@ -1978,6 +2728,36 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. + "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -2187,6 +2967,70 @@Method Details
}, }, }, + "dsseAttestation": { # An occurrence describing an attestation on a resource # This represents a DSSE attestation occurrence + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # If doing something security critical, make sure to verify the signatures in this metadata. + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, + "statement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". + "predicateType": "A String", # "https://in-toto.io/Provenance/v0.1" for InTotoProvenance. + "provenance": { + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ + { + "digest": { # "": "" + "a_key": "A String", + }, + "name": "A String", + }, + ], + "type": "A String", # Always "https://in-toto.io/Statement/v0.1". + }, + }, + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # https://github.com/secure-systems-lab/dsse + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, "installation": { # This represents how a particular software package may be installed on a system. # Describes the installation of a package on the linked resource. "location": [ # All of the places within the filesystem versions of this package have been found. { # An occurrence of a particular package installation found within a system's filesystem. e.g. glibc was found in /var/lib/dpkg/status @@ -2216,6 +3060,52 @@Method Details
"uri": "A String", # The unique URI of the resource. E.g., "https://gcr.io/project/image@sha256:foo" for a Docker image. }, "resourceUrl": "A String", # The unique URL of the image or the container for which the `Occurrence` applies. For example, https://gcr.io/project/image@sha256:foo This field can be used as a filter in list requests. + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific software bill of materials document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific relationship between SPDX elements. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this `Occurrence` was last updated. "upgrade": { # An Upgrade Occurrence represents that a specific resource_url could install a specific upgrade. This presence is supplied via local sources (i.e. it is present in the mirror and the running system has noticed its availability). # Describes an upgrade. "distribution": { # The Upgrade Distribution represents metadata about the Upgrade for each operating system (CPE). Some distributions have additional metadata around updates, classifying them into various categories and severities. # Metadata about the upgrade for available for the specific operating system for the resource_url. This allows efficient filtering, as well as making it easier to use the occurrence. diff --git a/docs/dyn/containeranalysis_v1alpha1.providers.notes.html b/docs/dyn/containeranalysis_v1alpha1.providers.notes.html index 6392800ae94..10154d07c11 100644 --- a/docs/dyn/containeranalysis_v1alpha1.providers.notes.html +++ b/docs/dyn/containeranalysis_v1alpha1.providers.notes.html @@ -175,6 +175,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -204,7 +209,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. @@ -330,6 +375,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -359,7 +409,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. @@ -508,6 +598,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -537,7 +632,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. @@ -718,6 +853,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -747,7 +887,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. @@ -889,6 +1069,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -918,7 +1103,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. @@ -1043,6 +1268,11 @@Method Details
"discovery": { # A note that indicates a type of analysis a provider would perform. This note exists in a provider's project. A `Discovery` occurrence is created in a consumer's project at the start of analysis. The occurrence's operation will indicate the status of the analysis. Absence of an occurrence linked to this note for a resource indicates that analysis hasn't started. # A note describing a provider/analysis type. "analysisKind": "A String", # The kind of analysis that is handled by this discovery. }, + "dsseAttestation": { # A note describing an attestation # A note describing a dsse attestation note. + "hint": { # This submessage provides human-readable hints about the purpose of the authority. Because the name of a note acts as its resource reference, it is important to disambiguate the canonical name of the Note (which might be a UUID for security purposes) from "readable" names more suitable for debug output. Note that these hints should not be used to look up authorities in security sensitive contexts, such as when looking up attestations to verify. # DSSEHint hints at the purpose of the attestation authority. + "humanReadableName": "A String", # Required. The human readable name of this attestation authority, for example "cloudbuild-prod". + }, + }, "expirationTime": "A String", # Time of expiration for this note, null if note does not expire. "kind": "A String", # Output only. This explicitly denotes which kind of note is specified. This field can be used as a filter in list requests. "longDescription": "A String", # A detailed description of this `Note`. @@ -1072,7 +1302,47 @@Method Details
"url": "A String", # Specific URL to associate with the note }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing a software bill of materials. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this `Note`. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing an SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing an SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing a relationship between SPDX elements. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "upgrade": { # An Upgrade Note represents a potential upgrade of a package to a given version. For each package version combination (i.e. bash 4.0, bash 4.1, bash 4.1.2), there will be a Upgrade Note. # A note describing an upgrade. "distributions": [ # Metadata about the upgrade for each specific operating system. diff --git a/docs/dyn/containeranalysis_v1alpha1.providers.notes.occurrences.html b/docs/dyn/containeranalysis_v1alpha1.providers.notes.occurrences.html index f59bd12ffe6..6b7155c8194 100644 --- a/docs/dyn/containeranalysis_v1alpha1.providers.notes.occurrences.html +++ b/docs/dyn/containeranalysis_v1alpha1.providers.notes.occurrences.html @@ -118,6 +118,36 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. + "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -327,6 +357,70 @@Method Details
}, }, }, + "dsseAttestation": { # An occurrence describing an attestation on a resource # This represents a DSSE attestation occurrence + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # If doing something security critical, make sure to verify the signatures in this metadata. + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, + "statement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". + "predicateType": "A String", # "https://in-toto.io/Provenance/v0.1" for InTotoProvenance. + "provenance": { + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + "A String", + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + "a_key": "A String", + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ + { + "digest": { # "": "" + "a_key": "A String", + }, + "name": "A String", + }, + ], + "type": "A String", # Always "https://in-toto.io/Statement/v0.1". + }, + }, + "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # https://github.com/secure-systems-lab/dsse + "payload": "A String", # The bytes being signed + "payloadType": "A String", # The type of payload being signed + "signatures": [ # The signatures over the payload + { # A DSSE signature + "keyid": "A String", # A reference id to the key being used for signing + "sig": "A String", # The signature itself + }, + ], + }, "installation": { # This represents how a particular software package may be installed on a system. # Describes the installation of a package on the linked resource. "location": [ # All of the places within the filesystem versions of this package have been found. { # An occurrence of a particular package installation found within a system's filesystem. e.g. glibc was found in /var/lib/dpkg/status @@ -356,6 +450,52 @@Method Details
"uri": "A String", # The unique URI of the resource. E.g., "https://gcr.io/project/image@sha256:foo" for a Docker image. }, "resourceUrl": "A String", # The unique URL of the image or the container for which the `Occurrence` applies. For example, https://gcr.io/project/image@sha256:foo This field can be used as a filter in list requests. + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific software bill of materials document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific relationship between SPDX elements. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this `Occurrence` was last updated. "upgrade": { # An Upgrade Occurrence represents that a specific resource_url could install a specific upgrade. This presence is supplied via local sources (i.e. it is present in the mirror and the running system has noticed its availability). # Describes an upgrade. "distribution": { # The Upgrade Distribution represents metadata about the Upgrade for each operating system (CPE). Some distributions have additional metadata around updates, classifying them into various categories and severities. # Metadata about the upgrade for available for the specific operating system for the resource_url. This allows efficient filtering, as well as making it easier to use the occurrence. diff --git a/docs/dyn/containeranalysis_v1beta1.projects.notes.html b/docs/dyn/containeranalysis_v1beta1.projects.notes.html index 7ae8d500959..531a5100ded 100644 --- a/docs/dyn/containeranalysis_v1beta1.projects.notes.html +++ b/docs/dyn/containeranalysis_v1beta1.projects.notes.html @@ -219,7 +219,47 @@Method Details
"url": "A String", # Specific URL associated with the resource. }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing SPDX Document which represents SBOM. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this note. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing SPDX Relationship. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "vulnerability": { # Vulnerability provides metadata about a security vulnerability in a Note. # A note describing a package vulnerability. "cvssScore": 3.14, # The CVSS score for this vulnerability. @@ -399,7 +439,47 @@Method Details
"url": "A String", # Specific URL associated with the resource. }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing SPDX Document which represents SBOM. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this note. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing SPDX Relationship. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "vulnerability": { # Vulnerability provides metadata about a security vulnerability in a Note. # A note describing a package vulnerability. "cvssScore": 3.14, # The CVSS score for this vulnerability. @@ -584,7 +664,47 @@Method Details
"url": "A String", # Specific URL associated with the resource. }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing SPDX Document which represents SBOM. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this note. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing SPDX Relationship. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "vulnerability": { # Vulnerability provides metadata about a security vulnerability in a Note. # A note describing a package vulnerability. "cvssScore": 3.14, # The CVSS score for this vulnerability. @@ -761,7 +881,47 @@Method Details
"url": "A String", # Specific URL associated with the resource. }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing SPDX Document which represents SBOM. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this note. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing SPDX Relationship. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "vulnerability": { # Vulnerability provides metadata about a security vulnerability in a Note. # A note describing a package vulnerability. "cvssScore": 3.14, # The CVSS score for this vulnerability. @@ -962,7 +1122,47 @@Method Details
"url": "A String", # Specific URL associated with the resource. }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing SPDX Document which represents SBOM. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this note. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing SPDX Relationship. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "vulnerability": { # Vulnerability provides metadata about a security vulnerability in a Note. # A note describing a package vulnerability. "cvssScore": 3.14, # The CVSS score for this vulnerability. @@ -1194,7 +1394,47 @@Method Details
"url": "A String", # Specific URL associated with the resource. }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing SPDX Document which represents SBOM. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this note. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing SPDX Relationship. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "vulnerability": { # Vulnerability provides metadata about a security vulnerability in a Note. # A note describing a package vulnerability. "cvssScore": 3.14, # The CVSS score for this vulnerability. @@ -1388,7 +1628,47 @@Method Details
"url": "A String", # Specific URL associated with the resource. }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing SPDX Document which represents SBOM. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this note. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing SPDX Relationship. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "vulnerability": { # Vulnerability provides metadata about a security vulnerability in a Note. # A note describing a package vulnerability. "cvssScore": 3.14, # The CVSS score for this vulnerability. @@ -1565,7 +1845,47 @@Method Details
"url": "A String", # Specific URL associated with the resource. }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing SPDX Document which represents SBOM. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this note. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing SPDX Relationship. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "vulnerability": { # Vulnerability provides metadata about a security vulnerability in a Note. # A note describing a package vulnerability. "cvssScore": 3.14, # The CVSS score for this vulnerability. diff --git a/docs/dyn/containeranalysis_v1beta1.projects.notes.occurrences.html b/docs/dyn/containeranalysis_v1beta1.projects.notes.occurrences.html index 5d18f6d0c80..78aae529b9b 100644 --- a/docs/dyn/containeranalysis_v1beta1.projects.notes.occurrences.html +++ b/docs/dyn/containeranalysis_v1beta1.projects.notes.occurrences.html @@ -368,6 +368,52 @@Method Details
"name": "A String", # Deprecated, do not use. Use uri instead. The name of the resource. For example, the name of a Docker image - "Debian". "uri": "A String", # Required. The unique URI of the resource. For example, `https://gcr.io/project/image@sha256:foo` for a Docker image. }, + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific SPDX Document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific SPDX Relationship. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this occurrence was last updated. "vulnerability": { # Details of a vulnerability Occurrence. # Describes a security vulnerability. "cvssScore": 3.14, # Output only. The CVSS score of this vulnerability. CVSS score is on a scale of 0-10 where 0 indicates low severity and 10 indicates high severity. diff --git a/docs/dyn/containeranalysis_v1beta1.projects.occurrences.html b/docs/dyn/containeranalysis_v1beta1.projects.occurrences.html index f42be6329f7..82f6fe08b91 100644 --- a/docs/dyn/containeranalysis_v1beta1.projects.occurrences.html +++ b/docs/dyn/containeranalysis_v1beta1.projects.occurrences.html @@ -384,6 +384,52 @@Method Details
"name": "A String", # Deprecated, do not use. Use uri instead. The name of the resource. For example, the name of a Docker image - "Debian". "uri": "A String", # Required. The unique URI of the resource. For example, `https://gcr.io/project/image@sha256:foo` for a Docker image. }, + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific SPDX Document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific SPDX Relationship. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this occurrence was last updated. "vulnerability": { # Details of a vulnerability Occurrence. # Describes a security vulnerability. "cvssScore": 3.14, # Output only. The CVSS score of this vulnerability. CVSS score is on a scale of 0-10 where 0 indicates low severity and 10 indicates high severity. @@ -699,6 +745,52 @@Method Details
"name": "A String", # Deprecated, do not use. Use uri instead. The name of the resource. For example, the name of a Docker image - "Debian". "uri": "A String", # Required. The unique URI of the resource. For example, `https://gcr.io/project/image@sha256:foo` for a Docker image. }, + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific SPDX Document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific SPDX Relationship. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this occurrence was last updated. "vulnerability": { # Details of a vulnerability Occurrence. # Describes a security vulnerability. "cvssScore": 3.14, # Output only. The CVSS score of this vulnerability. CVSS score is on a scale of 0-10 where 0 indicates low severity and 10 indicates high severity. @@ -1019,6 +1111,52 @@Method Details
"name": "A String", # Deprecated, do not use. Use uri instead. The name of the resource. For example, the name of a Docker image - "Debian". "uri": "A String", # Required. The unique URI of the resource. For example, `https://gcr.io/project/image@sha256:foo` for a Docker image. }, + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific SPDX Document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific SPDX Relationship. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this occurrence was last updated. "vulnerability": { # Details of a vulnerability Occurrence. # Describes a security vulnerability. "cvssScore": 3.14, # Output only. The CVSS score of this vulnerability. CVSS score is on a scale of 0-10 where 0 indicates low severity and 10 indicates high severity. @@ -1330,6 +1468,52 @@Method Details
"name": "A String", # Deprecated, do not use. Use uri instead. The name of the resource. For example, the name of a Docker image - "Debian". "uri": "A String", # Required. The unique URI of the resource. For example, `https://gcr.io/project/image@sha256:foo` for a Docker image. }, + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific SPDX Document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific SPDX Relationship. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this occurrence was last updated. "vulnerability": { # Details of a vulnerability Occurrence. # Describes a security vulnerability. "cvssScore": 3.14, # Output only. The CVSS score of this vulnerability. CVSS score is on a scale of 0-10 where 0 indicates low severity and 10 indicates high severity. @@ -1666,6 +1850,52 @@Method Details
"name": "A String", # Deprecated, do not use. Use uri instead. The name of the resource. For example, the name of a Docker image - "Debian". "uri": "A String", # Required. The unique URI of the resource. For example, `https://gcr.io/project/image@sha256:foo` for a Docker image. }, + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific SPDX Document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific SPDX Relationship. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this occurrence was last updated. "vulnerability": { # Details of a vulnerability Occurrence. # Describes a security vulnerability. "cvssScore": 3.14, # Output only. The CVSS score of this vulnerability. CVSS score is on a scale of 0-10 where 0 indicates low severity and 10 indicates high severity. @@ -1863,7 +2093,47 @@Method Details
"url": "A String", # Specific URL associated with the resource. }, ], + "sbom": { # DocumentNote represents an SPDX Document Creation Infromation section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # A note describing SPDX Document which represents SBOM. + "dataLicence": "A String", # Compliance with the SPDX specification includes populating the SPDX fields therein with data related to such fields ("SPDX-Metadata") + "spdxVersion": "A String", # Provide a reference number that can be used to understand how to parse and interpret the rest of the file + }, "shortDescription": "A String", # A one sentence description of this note. + "spdxFile": { # FileNote represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # A note describing SPDX File. + "checksum": [ # Provide a unique identifier to match analysis information on each specific file in a package + "A String", + ], + "fileType": "A String", # This field provides information about the type of file identified + "title": "A String", # Identify the full path and filename that corresponds to the file information in this section + }, + "spdxPackage": { # PackageNote represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # A note describing SPDX Package. + "analyzed": True or False, # Indicates whether the file content of this package has been available for or subjected to analysis when creating the SPDX document + "attribution": "A String", # A place for the SPDX data creator to record, at the package level, acknowledgements that may be needed to be communicated in some contexts + "checksum": "A String", # Provide an independently reproducible mechanism that permits unique identification of a specific package that correlates to the data in this SPDX file + "copyright": "A String", # Identify the copyright holders of the package, as well as any dates present + "detailedDescription": "A String", # A more detailed description of the package + "downloadLocation": "A String", # This section identifies the download Universal Resource Locator (URL), or a specific location within a version control system (VCS) for the package at the time that the SPDX file was created + "externalRefs": [ # ExternalRef + { # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "category": "A String", # An External Reference allows a Package to reference an external source of additional information, metadata, enumerations, asset identifiers, or downloadable content believed to be relevant to the Package + "comment": "A String", # Human-readable information about the purpose and target of the reference + "locator": "A String", # The unique string with no spaces necessary to access the package-specific information, metadata, or content within the target location + "type": "A String", # Type of category (e.g. 'npm' for the PACKAGE_MANAGER category) + }, + ], + "filesLicenseInfo": [ # Contain the license the SPDX file creator has concluded as governing the This field is to contain a list of all licenses found in the package. The relationship between licenses (i.e., conjunctive, disjunctive) is not specified in this field – it is simply a listing of all licenses found + "A String", + ], + "homePage": "A String", # Provide a place for the SPDX file creator to record a web site that serves as the package's home page + "licenseDeclared": "A String", # List the licenses that have been declared by the authors of the package + "originator": "A String", # If the package identified in the SPDX file originated from a different person or organization than identified as Package Supplier, this field identifies from where or whom the package originally came + "summaryDescription": "A String", # A short description of the package + "supplier": "A String", # Identify the actual distribution source for the package/directory identified in the SPDX file + "title": "A String", # Identify the full name of the package as given by the Package Originator + "verificationCode": "A String", # This field provides an independently reproducible mechanism identifying specific contents of a package based on the actual files (except the SPDX file itself, if it is included in the package) that make up each package and that correlates to the data in this SPDX file + "version": "A String", # Identify the version of the package + }, + "spdxRelationship": { # RelationshipNote represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # A note describing SPDX Relationship. + }, "updateTime": "A String", # Output only. The time this note was last updated. This field can be used as a filter in list requests. "vulnerability": { # Vulnerability provides metadata about a security vulnerability in a Note. # A note describing a package vulnerability. "cvssScore": 3.14, # The CVSS score for this vulnerability. @@ -2250,6 +2520,52 @@Method Details
"name": "A String", # Deprecated, do not use. Use uri instead. The name of the resource. For example, the name of a Docker image - "Debian". "uri": "A String", # Required. The unique URI of the resource. For example, `https://gcr.io/project/image@sha256:foo` for a Docker image. }, + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific SPDX Document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific SPDX Relationship. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this occurrence was last updated. "vulnerability": { # Details of a vulnerability Occurrence. # Describes a security vulnerability. "cvssScore": 3.14, # Output only. The CVSS score of this vulnerability. CVSS score is on a scale of 0-10 where 0 indicates low severity and 10 indicates high severity. @@ -2579,6 +2895,52 @@Method Details
"name": "A String", # Deprecated, do not use. Use uri instead. The name of the resource. For example, the name of a Docker image - "Debian". "uri": "A String", # Required. The unique URI of the resource. For example, `https://gcr.io/project/image@sha256:foo` for a Docker image. }, + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific SPDX Document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific SPDX Relationship. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this occurrence was last updated. "vulnerability": { # Details of a vulnerability Occurrence. # Describes a security vulnerability. "cvssScore": 3.14, # Output only. The CVSS score of this vulnerability. CVSS score is on a scale of 0-10 where 0 indicates low severity and 10 indicates high severity. @@ -2891,6 +3253,52 @@Method Details
"name": "A String", # Deprecated, do not use. Use uri instead. The name of the resource. For example, the name of a Docker image - "Debian". "uri": "A String", # Required. The unique URI of the resource. For example, `https://gcr.io/project/image@sha256:foo` for a Docker image. }, + "sbom": { # DocumentOccurrence represents an SPDX Document Creation Information section: https://spdx.github.io/spdx-spec/2-document-creation-information/ # Describes a specific SPDX Document. + "createTime": "A String", # Identify when the SPDX file was originally created. The date is to be specified according to combined date and time in UTC format as specified in ISO 8601 standard + "creatorComment": "A String", # A field for creators of the SPDX file to provide general comments about the creation of the SPDX file or any other relevant comment not included in the other fields + "creators": [ # Identify who (or what, in the case of a tool) created the SPDX file. If the SPDX file was created by an individual, indicate the person's name + "A String", + ], + "documentComment": "A String", # A field for creators of the SPDX file content to provide comments to the consumers of the SPDX document + "externalDocumentRefs": [ # Identify any external SPDX documents referenced within this SPDX document + "A String", + ], + "id": "A String", # Identify the current SPDX document which may be referenced in relationships by other files, packages internally and documents externally + "licenseListVersion": "A String", # A field for creators of the SPDX file to provide the version of the SPDX License List used when the SPDX file was created + "namespace": "A String", # Provide an SPDX document specific namespace as a unique absolute Uniform Resource Identifier (URI) as specified in RFC-3986, with the exception of the ‘#’ delimiter + "title": "A String", # Identify name of this document as designated by creator + }, + "spdxFile": { # FileOccurrence represents an SPDX File Information section: https://spdx.github.io/spdx-spec/4-file-information/ # Describes a specific SPDX File. + "attributions": [ # This field provides a place for the SPDX data creator to record, at the file level, acknowledgements that may be needed to be communicated in some contexts + "A String", + ], + "comment": "A String", # This field provides a place for the SPDX file creator to record any general comments about the file + "contributors": [ # This field provides a place for the SPDX file creator to record file contributors + "A String", + ], + "copyright": "A String", # Identify the copyright holder of the file, as well as any dates present + "filesLicenseInfo": [ # This field contains the license information actually found in the file, if any + "A String", + ], + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background references or analysis that went in to arriving at the Concluded License for a file + "licenseConcluded": "A String", # This field contains the license the SPDX file creator has concluded as governing the file or alternative values if the governing license cannot be determined + "notice": "A String", # This field provides a place for the SPDX file creator to record license notices or other such related notices found in the file + }, + "spdxPackage": { # PackageOccurrence represents an SPDX Package Information section: https://spdx.github.io/spdx-spec/3-package-information/ # Describes a specific SPDX Package. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the package being described + "filename": "A String", # Provide the actual file name of the package, or path of the directory being treated as a package + "id": "A String", # Uniquely identify any element in an SPDX document which may be referenced by other elements + "licenseComments": "A String", # This field provides a place for the SPDX file creator to record any relevant background information or analysis that went in to arriving at the Concluded License for a package + "licenseConcluded": "A String", # package or alternative values, if the governing license cannot be determined + "sourceInfo": "A String", # Provide a place for the SPDX file creator to record any relevant background information or additional comments about the origin of the package + }, + "spdxRelationship": { # RelationshipOccurrence represents an SPDX Relationship section: https://spdx.github.io/spdx-spec/7-relationships-between-SPDX-elements/ # Describes a specific SPDX Relationship. + "comment": "A String", # A place for the SPDX file creator to record any general comments about the relationship + "source": "A String", # Also referred to as SPDXRef-A The source SPDX element (file, package, etc) + "target": "A String", # Also referred to as SPDXRef-B The target SPDC element (file, package, etc) In cases where there are "known unknowns", the use of the keyword NOASSERTION can be used The keywords NONE can be used to indicate that an SPDX element (package/file/snippet) has no other elements connected by some relationship to it + "type": "A String", # The type of relationship between the source and target SPDX elements + }, "updateTime": "A String", # Output only. The time this occurrence was last updated. "vulnerability": { # Details of a vulnerability Occurrence. # Describes a security vulnerability. "cvssScore": 3.14, # Output only. The CVSS score of this vulnerability. CVSS score is on a scale of 0-10 where 0 indicates low severity and 10 indicates high severity. diff --git a/docs/dyn/content_v2.shippingsettings.html b/docs/dyn/content_v2.shippingsettings.html index a116eb2ee3b..55616ac6c54 100644 --- a/docs/dyn/content_v2.shippingsettings.html +++ b/docs/dyn/content_v2.shippingsettings.html @@ -148,7 +148,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -166,7 +166,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -461,7 +461,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -479,7 +479,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -766,7 +766,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -784,7 +784,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1158,7 +1158,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1176,7 +1176,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1470,7 +1470,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1488,7 +1488,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1764,7 +1764,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1782,7 +1782,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, diff --git a/docs/dyn/content_v2_1.products.html b/docs/dyn/content_v2_1.products.html index ea13daf023b..f1c1c495833 100644 --- a/docs/dyn/content_v2_1.products.html +++ b/docs/dyn/content_v2_1.products.html @@ -119,7 +119,7 @@Method Details
"feedId": "A String", # The Content API Supplemental Feed ID. If present then product insertion or deletion applies to a supplemental feed instead of primary Content API feed. "merchantId": "A String", # The ID of the managing account. "method": "A String", # The method of the batch entry. Acceptable values are: - "`delete`" - "`get`" - "`insert`" - "`update`" - "product": { # Required product attributes are primarily defined by the products data specification. See the Products Data Specification Help Center article for information. Product data. After inserting, updating, or deleting a product, it may take several minutes before changes take effect. # The product to insert. Only required if the method is `insert`. + "product": { # Required product attributes are primarily defined by the products data specification. See the Products Data Specification Help Center article for information. Product data. After inserting, updating, or deleting a product, it may take several minutes before changes take effect. # The product to insert or update. Only required if the method is `insert` or `update`. If the `update` method is used with `updateMask` only to delete a field, then this isn't required. For example, setting `salePrice` on the `updateMask` and not providing a `product` will result in an existing sale price on the product specified by `productId` being deleted. "additionalImageLinks": [ # Additional URLs of images of the item. "A String", ], @@ -310,7 +310,7 @@Method Details
"value": 3.14, # The measure of an item. }, }, - "productId": "A String", # The ID of the product to get or delete. Only defined if the method is `get` or `delete`. + "productId": "A String", # The ID of the product to mutate. Only defined if the method is `get`, `delete`, or `update`. or `delete`. "updateMask": "A String", # The comma-separated list of product attributes to be updated. Example: `"title,salePrice"`. Attributes specified in the update mask without a value specified in the body will be deleted from the product. Only top-level product attributes can be updated. If not defined, product attributes with set values will be updated and other attributes will stay unchanged. Only defined if the method is `update`. }, ], diff --git a/docs/dyn/content_v2_1.shippingsettings.html b/docs/dyn/content_v2_1.shippingsettings.html index 5623448faee..85631d2208e 100644 --- a/docs/dyn/content_v2_1.shippingsettings.html +++ b/docs/dyn/content_v2_1.shippingsettings.html @@ -148,7 +148,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -166,7 +166,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -460,7 +460,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -478,7 +478,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -765,7 +765,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -783,7 +783,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1157,7 +1157,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1175,7 +1175,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1469,7 +1469,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1487,7 +1487,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1762,7 +1762,7 @@Method Details
"timezone": "A String", # Timezone identifier for the cutoff time. A list of identifiers can be found in the AdWords API documentation. E.g. "Europe/Zurich". Required. }, "handlingBusinessDayConfig": { # The business days during which orders can be handled. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, @@ -1780,7 +1780,7 @@Method Details
"minHandlingTimeInDays": 42, # Minimum number of business days spent before an order is shipped. 0 means same day shipped, 1 means next day shipped. "minTransitTimeInDays": 42, # Minimum number of business days that is spent in transit. 0 means same day delivery, 1 means next day delivery. Either `{min,max}TransitTimeInDays` or `transitTimeTable` must be set, but not both. "transitBusinessDayConfig": { # The business days during which orders can be in-transit. If not provided, Monday to Friday business days will be assumed. - "businessDays": [ # Regular business days. May not be empty. + "businessDays": [ # Regular business days, such as '"monday"'. May not be empty. "A String", ], }, diff --git a/docs/dyn/datacatalog_v1.entries.html b/docs/dyn/datacatalog_v1.entries.html index 14880637987..3b90eaf94d8 100644 --- a/docs/dyn/datacatalog_v1.entries.html +++ b/docs/dyn/datacatalog_v1.entries.html @@ -93,7 +93,7 @@Method Details
Args: fullyQualifiedName: string, Fully qualified name (FQN) of the resource. FQNs take two forms: * For non-regionalized resources: `{SYSTEM}:{PROJECT}.{PATH_TO_RESOURCE_SEPARATED_WITH_DOTS}` * For regionalized resources: `{SYSTEM}:{PROJECT}.{LOCATION_ID}.{PATH_TO_RESOURCE_SEPARATED_WITH_DOTS}` Example for a DPMS table: `dataproc_metastore:{PROJECT_ID}.{LOCATION_ID}.{INSTANCE_ID}.{DATABASE_ID}.{TABLE_ID}` linkedResource: string, The full name of the Google Cloud Platform resource the Data Catalog entry represents. For more information, see [Full Resource Name] (https://cloud.google.com/apis/design/resource_names#full_resource_name). Full names are case-sensitive. For example: * `//bigquery.googleapis.com/projects/{PROJECT_ID}/datasets/{DATASET_ID}/tables/{TABLE_ID}` * `//pubsub.googleapis.com/projects/{PROJECT_ID}/topics/{TOPIC_ID}` - sqlResource: string, The SQL name of the entry. SQL names are case-sensitive. Examples: * `pubsub.{PROJECT_ID}.{TOPIC_ID}` * `pubsub.{PROJECT_ID}.{TOPIC.ID.SEPARATED.WITH.DOTS}` * `bigquery.table.{PROJECT_ID}.{DATASET_ID}.{TABLE_ID}` * `bigquery.dataset.{PROJECT_ID}.{DATASET_ID}` * `datacatalog.entry.{PROJECT_ID}.{LOCATION_ID}.{ENTRY_GROUP_ID}.{ENTRY_ID}` Identifiers (`*_ID`) should comply with the [Lexical structure in Standard SQL] (https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical). + sqlResource: string, The SQL name of the entry. SQL names are case-sensitive. Examples: * `pubsub.topic.{PROJECT_ID}.{TOPIC_ID}` * `pubsub.topic.{PROJECT_ID}.`\``{TOPIC.ID.SEPARATED.WITH.DOTS}`\` * `bigquery.table.{PROJECT_ID}.{DATASET_ID}.{TABLE_ID}` * `bigquery.dataset.{PROJECT_ID}.{DATASET_ID}` * `datacatalog.entry.{PROJECT_ID}.{LOCATION_ID}.{ENTRY_GROUP_ID}.{ENTRY_ID}` Identifiers (`*_ID`) should comply with the [Lexical structure in Standard SQL] (https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical). x__xgafv: string, V1 error format. Allowed values 1 - v1 error format diff --git a/docs/dyn/datacatalog_v1.projects.locations.tagTemplates.html b/docs/dyn/datacatalog_v1.projects.locations.tagTemplates.html index 34d756f8d15..56877bfa4a3 100644 --- a/docs/dyn/datacatalog_v1.projects.locations.tagTemplates.html +++ b/docs/dyn/datacatalog_v1.projects.locations.tagTemplates.html @@ -139,7 +139,7 @@Method Details
}, }, }, - "isPubliclyReadable": True or False, # Indicates whether this is a public tag template. Every user has view access to a *public* tag template by default. This means that: * Every user can use this tag template to tag an entry. * If an entry is tagged using the tag template, the tag is always shown in the response to ``ListTags`` called on the entry. * To get the template using the GetTagTemplate method, you need view access either on the project or the organization the tag template resides in but no other permission is needed. * Operations on the tag template other than viewing (for example, editing IAM policies) follow standard IAM structures. Tags created with a public tag template are referred to as public tags. You can search for a public tag by value with a simple search query instead of using a ``tag:`` predicate. Public tag templates may not appear in search results depending on scope, see: include_public_tag_templates + "isPubliclyReadable": True or False, # Indicates whether this is a public tag template. Every user has view access to a *public* tag template by default. This means that: * Every user can use this tag template to tag an entry. * If an entry is tagged using the tag template, the tag is always shown in the response to ``ListTags`` called on the entry. * To get the template using the GetTagTemplate method, you need view access either on the project or the organization the tag template resides in but no other permission is needed. * Operations on the tag template other than viewing (for example, editing IAM policies) follow standard IAM structures. Tags created with a public tag template are referred to as public tags. You can search for a public tag by value with a simple search query instead of using a ``tag:`` predicate. Public tag templates may not appear in search results depending on scope, see: include_public_tag_templates Note: If an [IAM domain restriction](https://cloud.google.com/resource-manager/docs/organization-policy/restricting-domains) is configured in the tag template's location, the public access will not be enabled but the simple search for tag values will still work. "name": "A String", # The resource name of the tag template in URL format. Note: The tag template itself and its child resources might not be stored in the location specified in its name. } @@ -173,7 +173,7 @@Method Details
}, }, }, - "isPubliclyReadable": True or False, # Indicates whether this is a public tag template. Every user has view access to a *public* tag template by default. This means that: * Every user can use this tag template to tag an entry. * If an entry is tagged using the tag template, the tag is always shown in the response to ``ListTags`` called on the entry. * To get the template using the GetTagTemplate method, you need view access either on the project or the organization the tag template resides in but no other permission is needed. * Operations on the tag template other than viewing (for example, editing IAM policies) follow standard IAM structures. Tags created with a public tag template are referred to as public tags. You can search for a public tag by value with a simple search query instead of using a ``tag:`` predicate. Public tag templates may not appear in search results depending on scope, see: include_public_tag_templates + "isPubliclyReadable": True or False, # Indicates whether this is a public tag template. Every user has view access to a *public* tag template by default. This means that: * Every user can use this tag template to tag an entry. * If an entry is tagged using the tag template, the tag is always shown in the response to ``ListTags`` called on the entry. * To get the template using the GetTagTemplate method, you need view access either on the project or the organization the tag template resides in but no other permission is needed. * Operations on the tag template other than viewing (for example, editing IAM policies) follow standard IAM structures. Tags created with a public tag template are referred to as public tags. You can search for a public tag by value with a simple search query instead of using a ``tag:`` predicate. Public tag templates may not appear in search results depending on scope, see: include_public_tag_templates Note: If an [IAM domain restriction](https://cloud.google.com/resource-manager/docs/organization-policy/restricting-domains) is configured in the tag template's location, the public access will not be enabled but the simple search for tag values will still work. "name": "A String", # The resource name of the tag template in URL format. Note: The tag template itself and its child resources might not be stored in the location specified in its name. }
Close httplib2 connections.
create(name, body=None, x__xgafv=None)
Create a `Release`. Release names should reflect the developer's deployment practices. For example, the release name may include the environment name, application name, application version, or any other name meaningful to the developer. Once a `Release` refers to a `Ruleset`, the rules can be enforced by Firebase Rules-enabled services. More than one `Release` may be 'live' concurrently. Consider the following three `Release` names for `projects/foo` and the `Ruleset` to which they refer. Release Name | Ruleset Name --------------------------------|------------- projects/foo/releases/prod | projects/foo/rulesets/uuid123 projects/foo/releases/prod/beta | projects/foo/rulesets/uuid123 projects/foo/releases/prod/v23 | projects/foo/rulesets/uuid456 The table reflects the `Ruleset` rollout in progress. The `prod` and `prod/beta` releases refer to the same `Ruleset`. However, `prod/v23` refers to a new `Ruleset`. The `Ruleset` reference for a `Release` may be updated using the UpdateRelease method.
+Create a `Release`. Release names should reflect the developer's deployment practices. For example, the release name may include the environment name, application name, application version, or any other name meaningful to the developer. Once a `Release` refers to a `Ruleset`, the rules can be enforced by Firebase Rules-enabled services. More than one `Release` may be 'live' concurrently. Consider the following three `Release` names for `projects/foo` and the `Ruleset` to which they refer. Release Name -> Ruleset Name * projects/foo/releases/prod -> projects/foo/rulesets/uuid123 * projects/foo/releases/prod/beta -> projects/foo/rulesets/uuid123 * projects/foo/releases/prod/v23 -> projects/foo/rulesets/uuid456 The relationships reflect a `Ruleset` rollout in progress. The `prod` and `prod/beta` releases refer to the same `Ruleset`. However, `prod/v23` refers to a new `Ruleset`. The `Ruleset` reference for a `Release` may be updated using the UpdateRelease method.
Delete a `Release` by resource name.
@@ -106,7 +106,7 @@create(name, body=None, x__xgafv=None)
- Create a `Release`. Release names should reflect the developer's deployment practices. For example, the release name may include the environment name, application name, application version, or any other name meaningful to the developer. Once a `Release` refers to a `Ruleset`, the rules can be enforced by Firebase Rules-enabled services. More than one `Release` may be 'live' concurrently. Consider the following three `Release` names for `projects/foo` and the `Ruleset` to which they refer. Release Name | Ruleset Name --------------------------------|------------- projects/foo/releases/prod | projects/foo/rulesets/uuid123 projects/foo/releases/prod/beta | projects/foo/rulesets/uuid123 projects/foo/releases/prod/v23 | projects/foo/rulesets/uuid456 The table reflects the `Ruleset` rollout in progress. The `prod` and `prod/beta` releases refer to the same `Ruleset`. However, `prod/v23` refers to a new `Ruleset`. The `Ruleset` reference for a `Release` may be updated using the UpdateRelease method.
+ Create a `Release`. Release names should reflect the developer's deployment practices. For example, the release name may include the environment name, application name, application version, or any other name meaningful to the developer. Once a `Release` refers to a `Ruleset`, the rules can be enforced by Firebase Rules-enabled services. More than one `Release` may be 'live' concurrently. Consider the following three `Release` names for `projects/foo` and the `Ruleset` to which they refer. Release Name -> Ruleset Name * projects/foo/releases/prod -> projects/foo/rulesets/uuid123 * projects/foo/releases/prod/beta -> projects/foo/rulesets/uuid123 * projects/foo/releases/prod/v23 -> projects/foo/rulesets/uuid456 The relationships reflect a `Ruleset` rollout in progress. The `prod` and `prod/beta` releases refer to the same `Ruleset`. However, `prod/v23` refers to a new `Ruleset`. The `Ruleset` reference for a `Release` may be updated using the UpdateRelease method.
Args:
name: string, Required. Resource name for the project which owns this `Release`. Format: `projects/{project_id}` (required)
diff --git a/docs/dyn/firestore_v1.projects.databases.documents.html b/docs/dyn/firestore_v1.projects.databases.documents.html
index 58873196aeb..02464f5b2f5 100644
--- a/docs/dyn/firestore_v1.projects.databases.documents.html
+++ b/docs/dyn/firestore_v1.projects.databases.documents.html
@@ -175,7 +175,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -230,31 +234,16 @@ Method Details
{ # A transformation of a field of the document.
"appendMissingElements": { # An array value. # Append the given elements in order if they are not already present in the current field value. If the field is not an array, or if the field does not yet exist, it is first set to the empty array. Equivalent numbers of different types (e.g. 3L and 3.0) are considered equal when checking if a value is missing. NaN is equal to NaN, and Null is equal to Null. If the input contains multiple equivalent values, only the first will be considered. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"fieldPath": "A String", # The path of the field. See Document.fields for the field path syntax reference.
"increment": { # A message that can hold any of the supported value types. # Adds the given value to the field's current value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If either of the given value or the current field value are doubles, both values will be interpreted as doubles. Double arithmetic and representation of double values follow IEEE 754 semantics. If there is positive/negative integer overflow, the field is resolved to the largest magnitude positive/negative integer.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -274,7 +263,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"maximum": { # A message that can hold any of the supported value types. # Sets the field to the maximum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If a maximum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the larger operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The maximum of a zero stored value and zero input value is always the stored value. The maximum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -294,7 +287,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"minimum": { # A message that can hold any of the supported value types. # Sets the field to the minimum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the input value. If a minimum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the smaller operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The minimum of a zero stored value and zero input value is always the stored value. The minimum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -315,26 +312,7 @@ Method Details
},
"removeAllFromArray": { # An array value. # Remove all of the given elements from the array in the field. If the field is not an array, or if the field does not yet exist, it is set to the empty array. Equivalent numbers of the different types (e.g. 3L and 3.0) are considered equal when deciding whether an element should be removed. NaN is equal to NaN, and Null is equal to Null. This will remove all equivalent values if there are duplicates. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"setToServerValue": "A String", # Sets the field to the given server value.
@@ -345,7 +323,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -377,31 +359,16 @@ Method Details
{ # A transformation of a field of the document.
"appendMissingElements": { # An array value. # Append the given elements in order if they are not already present in the current field value. If the field is not an array, or if the field does not yet exist, it is first set to the empty array. Equivalent numbers of different types (e.g. 3L and 3.0) are considered equal when checking if a value is missing. NaN is equal to NaN, and Null is equal to Null. If the input contains multiple equivalent values, only the first will be considered. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"fieldPath": "A String", # The path of the field. See Document.fields for the field path syntax reference.
"increment": { # A message that can hold any of the supported value types. # Adds the given value to the field's current value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If either of the given value or the current field value are doubles, both values will be interpreted as doubles. Double arithmetic and representation of double values follow IEEE 754 semantics. If there is positive/negative integer overflow, the field is resolved to the largest magnitude positive/negative integer.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -421,7 +388,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"maximum": { # A message that can hold any of the supported value types. # Sets the field to the maximum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If a maximum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the larger operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The maximum of a zero stored value and zero input value is always the stored value. The maximum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -441,7 +412,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"minimum": { # A message that can hold any of the supported value types. # Sets the field to the minimum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the input value. If a minimum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the smaller operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The minimum of a zero stored value and zero input value is always the stored value. The minimum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -462,26 +437,7 @@ Method Details
},
"removeAllFromArray": { # An array value. # Remove all of the given elements from the array in the field. If the field is not an array, or if the field does not yet exist, it is set to the empty array. Equivalent numbers of the different types (e.g. 3L and 3.0) are considered equal when deciding whether an element should be removed. NaN is equal to NaN, and Null is equal to Null. This will remove all equivalent values if there are duplicates. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"setToServerValue": "A String", # Sets the field to the given server value.
@@ -515,7 +471,11 @@ Method Details
{ # The result of applying a write.
"transformResults": [ # The results of applying each DocumentTransform.FieldTransform, in the same order.
{ # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -603,31 +563,16 @@ Method Details
{ # A transformation of a field of the document.
"appendMissingElements": { # An array value. # Append the given elements in order if they are not already present in the current field value. If the field is not an array, or if the field does not yet exist, it is first set to the empty array. Equivalent numbers of different types (e.g. 3L and 3.0) are considered equal when checking if a value is missing. NaN is equal to NaN, and Null is equal to Null. If the input contains multiple equivalent values, only the first will be considered. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"fieldPath": "A String", # The path of the field. See Document.fields for the field path syntax reference.
"increment": { # A message that can hold any of the supported value types. # Adds the given value to the field's current value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If either of the given value or the current field value are doubles, both values will be interpreted as doubles. Double arithmetic and representation of double values follow IEEE 754 semantics. If there is positive/negative integer overflow, the field is resolved to the largest magnitude positive/negative integer.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -647,7 +592,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"maximum": { # A message that can hold any of the supported value types. # Sets the field to the maximum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If a maximum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the larger operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The maximum of a zero stored value and zero input value is always the stored value. The maximum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -667,7 +616,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"minimum": { # A message that can hold any of the supported value types. # Sets the field to the minimum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the input value. If a minimum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the smaller operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The minimum of a zero stored value and zero input value is always the stored value. The minimum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -688,26 +641,7 @@ Method Details
},
"removeAllFromArray": { # An array value. # Remove all of the given elements from the array in the field. If the field is not an array, or if the field does not yet exist, it is set to the empty array. Equivalent numbers of the different types (e.g. 3L and 3.0) are considered equal when deciding whether an element should be removed. NaN is equal to NaN, and Null is equal to Null. This will remove all equivalent values if there are duplicates. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"setToServerValue": "A String", # Sets the field to the given server value.
@@ -718,7 +652,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -750,31 +688,16 @@ Method Details
{ # A transformation of a field of the document.
"appendMissingElements": { # An array value. # Append the given elements in order if they are not already present in the current field value. If the field is not an array, or if the field does not yet exist, it is first set to the empty array. Equivalent numbers of different types (e.g. 3L and 3.0) are considered equal when checking if a value is missing. NaN is equal to NaN, and Null is equal to Null. If the input contains multiple equivalent values, only the first will be considered. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"fieldPath": "A String", # The path of the field. See Document.fields for the field path syntax reference.
"increment": { # A message that can hold any of the supported value types. # Adds the given value to the field's current value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If either of the given value or the current field value are doubles, both values will be interpreted as doubles. Double arithmetic and representation of double values follow IEEE 754 semantics. If there is positive/negative integer overflow, the field is resolved to the largest magnitude positive/negative integer.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -794,7 +717,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"maximum": { # A message that can hold any of the supported value types. # Sets the field to the maximum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If a maximum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the larger operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The maximum of a zero stored value and zero input value is always the stored value. The maximum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -814,7 +741,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"minimum": { # A message that can hold any of the supported value types. # Sets the field to the minimum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the input value. If a minimum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the smaller operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The minimum of a zero stored value and zero input value is always the stored value. The minimum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -835,26 +766,7 @@ Method Details
},
"removeAllFromArray": { # An array value. # Remove all of the given elements from the array in the field. If the field is not an array, or if the field does not yet exist, it is set to the empty array. Equivalent numbers of the different types (e.g. 3L and 3.0) are considered equal when deciding whether an element should be removed. NaN is equal to NaN, and Null is equal to Null. This will remove all equivalent values if there are duplicates. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"setToServerValue": "A String", # Sets the field to the given server value.
@@ -878,7 +790,11 @@ Method Details
{ # The result of applying a write.
"transformResults": [ # The results of applying each DocumentTransform.FieldTransform, in the same order.
{ # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -918,7 +834,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -956,7 +876,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1022,7 +946,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1075,7 +1003,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1185,7 +1117,11 @@ Method Details
"before": True or False, # If the position is just before or just after the given values, relative to the sort order defined by the query.
"values": [ # The values that represent a position, in the order they appear in the order by clause of a query. Can contain fewer values than specified in the order by clause.
{ # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1233,7 +1169,11 @@ Method Details
"before": True or False, # If the position is just before or just after the given values, relative to the sort order defined by the query.
"values": [ # The values that represent a position, in the order they appear in the order by clause of a query. Can contain fewer values than specified in the order by clause.
{ # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1267,7 +1207,11 @@ Method Details
},
"op": "A String", # The operator to filter by.
"value": { # A message that can hold any of the supported value types. # The value to compare to.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1320,7 +1264,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1406,7 +1354,11 @@ Method Details
"before": True or False, # If the position is just before or just after the given values, relative to the sort order defined by the query.
"values": [ # The values that represent a position, in the order they appear in the order by clause of a query. Can contain fewer values than specified in the order by clause.
{ # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1454,7 +1406,11 @@ Method Details
"before": True or False, # If the position is just before or just after the given values, relative to the sort order defined by the query.
"values": [ # The values that represent a position, in the order they appear in the order by clause of a query. Can contain fewer values than specified in the order by clause.
{ # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1488,7 +1444,11 @@ Method Details
},
"op": "A String", # The operator to filter by.
"value": { # A message that can hold any of the supported value types. # The value to compare to.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1533,7 +1493,11 @@ Method Details
"before": True or False, # If the position is just before or just after the given values, relative to the sort order defined by the query.
"values": [ # The values that represent a position, in the order they appear in the order by clause of a query. Can contain fewer values than specified in the order by clause.
{ # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1585,7 +1549,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1625,7 +1593,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1699,7 +1671,11 @@ Method Details
"before": True or False, # If the position is just before or just after the given values, relative to the sort order defined by the query.
"values": [ # The values that represent a position, in the order they appear in the order by clause of a query. Can contain fewer values than specified in the order by clause.
{ # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1747,7 +1723,11 @@ Method Details
"before": True or False, # If the position is just before or just after the given values, relative to the sort order defined by the query.
"values": [ # The values that represent a position, in the order they appear in the order by clause of a query. Can contain fewer values than specified in the order by clause.
{ # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1781,7 +1761,11 @@ Method Details
},
"op": "A String", # The operator to filter by.
"value": { # A message that can hold any of the supported value types. # The value to compare to.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1825,7 +1809,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1882,31 +1870,16 @@ Method Details
{ # A transformation of a field of the document.
"appendMissingElements": { # An array value. # Append the given elements in order if they are not already present in the current field value. If the field is not an array, or if the field does not yet exist, it is first set to the empty array. Equivalent numbers of different types (e.g. 3L and 3.0) are considered equal when checking if a value is missing. NaN is equal to NaN, and Null is equal to Null. If the input contains multiple equivalent values, only the first will be considered. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"fieldPath": "A String", # The path of the field. See Document.fields for the field path syntax reference.
"increment": { # A message that can hold any of the supported value types. # Adds the given value to the field's current value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If either of the given value or the current field value are doubles, both values will be interpreted as doubles. Double arithmetic and representation of double values follow IEEE 754 semantics. If there is positive/negative integer overflow, the field is resolved to the largest magnitude positive/negative integer.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1926,7 +1899,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"maximum": { # A message that can hold any of the supported value types. # Sets the field to the maximum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If a maximum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the larger operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The maximum of a zero stored value and zero input value is always the stored value. The maximum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1946,7 +1923,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"minimum": { # A message that can hold any of the supported value types. # Sets the field to the minimum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the input value. If a minimum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the smaller operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The minimum of a zero stored value and zero input value is always the stored value. The minimum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -1967,26 +1948,7 @@ Method Details
},
"removeAllFromArray": { # An array value. # Remove all of the given elements from the array in the field. If the field is not an array, or if the field does not yet exist, it is set to the empty array. Equivalent numbers of the different types (e.g. 3L and 3.0) are considered equal when deciding whether an element should be removed. NaN is equal to NaN, and Null is equal to Null. This will remove all equivalent values if there are duplicates. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"setToServerValue": "A String", # Sets the field to the given server value.
@@ -1997,7 +1959,11 @@ Method Details
"createTime": "A String", # Output only. The time at which the document was created. This value increases monotonically when a document is deleted then recreated. It can also be compared to values from other documents and the `read_time` of a query.
"fields": { # The document's fields. The map keys represent field names. A simple field name contains only characters `a` to `z`, `A` to `Z`, `0` to `9`, or `_`, and must not start with `0` to `9`. For example, `foo_bar_17`. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty. Field paths may be used in other contexts to refer to structured fields defined here. For `map_value`, the field path is represented by the simple or quoted field names of the containing fields, delimited by `.`. For example, the structured field `"foo" : { map_value: { "x&y" : { string_value: "hello" }}}` would be represented by the field path `foo.x&y`. Within a field path, a quoted field name starts and ends with `` ` `` and may contain any character. Some characters, including `` ` ``, must be escaped using a `\`. For example, `` `x&y` `` represents `x&y` and `` `bak\`tik` `` represents `` bak`tik ``.
"a_key": { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -2029,31 +1995,16 @@ Method Details
{ # A transformation of a field of the document.
"appendMissingElements": { # An array value. # Append the given elements in order if they are not already present in the current field value. If the field is not an array, or if the field does not yet exist, it is first set to the empty array. Equivalent numbers of different types (e.g. 3L and 3.0) are considered equal when checking if a value is missing. NaN is equal to NaN, and Null is equal to Null. If the input contains multiple equivalent values, only the first will be considered. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"fieldPath": "A String", # The path of the field. See Document.fields for the field path syntax reference.
"increment": { # A message that can hold any of the supported value types. # Adds the given value to the field's current value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If either of the given value or the current field value are doubles, both values will be interpreted as doubles. Double arithmetic and representation of double values follow IEEE 754 semantics. If there is positive/negative integer overflow, the field is resolved to the largest magnitude positive/negative integer.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -2073,7 +2024,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"maximum": { # A message that can hold any of the supported value types. # Sets the field to the maximum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the given value. If a maximum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the larger operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The maximum of a zero stored value and zero input value is always the stored value. The maximum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -2093,7 +2048,11 @@ Method Details
"timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
},
"minimum": { # A message that can hold any of the supported value types. # Sets the field to the minimum of its current value and the given value. This must be an integer or a double value. If the field is not an integer or double, or if the field does not yet exist, the transformation will set the field to the input value. If a minimum operation is applied where the field and the input value are of mixed types (that is - one is an integer and one is a double) the field takes on the type of the smaller operand. If the operands are equivalent (e.g. 3 and 3.0), the field does not change. 0, 0.0, and -0.0 are all zero. The minimum of a zero stored value and zero input value is always the stored value. The minimum of any numeric value x and NaN is NaN.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
@@ -2114,26 +2073,7 @@ Method Details
},
"removeAllFromArray": { # An array value. # Remove all of the given elements from the array in the field. If the field is not an array, or if the field does not yet exist, it is set to the empty array. Equivalent numbers of the different types (e.g. 3L and 3.0) are considered equal when deciding whether an element should be removed. NaN is equal to NaN, and Null is equal to Null. This will remove all equivalent values if there are duplicates. The corresponding transform_result will be the null value.
"values": [ # Values in the array.
- { # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
- "booleanValue": True or False, # A boolean value.
- "bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
- "doubleValue": 3.14, # A double value.
- "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
- "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
- "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
- },
- "integerValue": "A String", # An integer value.
- "mapValue": { # A map value. # A map value.
- "fields": { # The map's fields. The map keys represent field names. Field names matching the regular expression `__.*__` are reserved. Reserved field names are forbidden except in certain documented contexts. The map keys, represented as UTF-8, must not exceed 1,500 bytes and cannot be empty.
- "a_key": # Object with schema name: Value
- },
- },
- "nullValue": "A String", # A null value.
- "referenceValue": "A String", # A reference to a document. For example: `projects/{project_id}/databases/{database_id}/documents/{document_path}`.
- "stringValue": "A String", # A string value. The string, represented as UTF-8, must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes of the UTF-8 representation are considered by queries.
- "timestampValue": "A String", # A timestamp value. Precise only to microseconds. When stored, any additional precision is rounded down.
- },
+ # Object with schema name: Value
],
},
"setToServerValue": "A String", # Sets the field to the given server value.
@@ -2159,7 +2099,11 @@ Method Details
{ # The result of applying a write.
"transformResults": [ # The results of applying each DocumentTransform.FieldTransform, in the same order.
{ # A message that can hold any of the supported value types.
- "arrayValue": # Object with schema name: ArrayValue # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "arrayValue": { # An array value. # An array value. Cannot directly contain another array value, though can contain an map which contains another array.
+ "values": [ # Values in the array.
+ # Object with schema name: Value
+ ],
+ },
"booleanValue": True or False, # A boolean value.
"bytesValue": "A String", # A bytes value. Must not exceed 1 MiB - 89 bytes. Only the first 1,500 bytes are considered by queries.
"doubleValue": 3.14, # A double value.
diff --git a/docs/dyn/iam_v1.projects.serviceAccounts.html b/docs/dyn/iam_v1.projects.serviceAccounts.html
index 1fc1949c4d0..6903bbce80b 100644
--- a/docs/dyn/iam_v1.projects.serviceAccounts.html
+++ b/docs/dyn/iam_v1.projects.serviceAccounts.html
@@ -534,7 +534,7 @@ Method Details
The object takes the form of:
{ # Deprecated. [Migrate to Service Account Credentials API](https://cloud.google.com/iam/help/credentials/migrate-api). The service account sign JWT request.
- "payload": "A String", # Required. Deprecated. [Migrate to Service Account Credentials API](https://cloud.google.com/iam/help/credentials/migrate-api). The JWT payload to sign. Must be a serialized JSON object that contains a JWT Claims Set. For example: `{"sub": "user@example.com", "iat": 313435}` If the JWT Claims Set contains an expiration time (`exp`) claim, it must be an integer timestamp that is not in the past and no more than 1 hour in the future. If the JWT Claims Set does not contain an expiration time (`exp`) claim, this claim is added automatically, with a timestamp that is 1 hour in the future.
+ "payload": "A String", # Required. Deprecated. [Migrate to Service Account Credentials API](https://cloud.google.com/iam/help/credentials/migrate-api). The JWT payload to sign. Must be a serialized JSON object that contains a JWT Claims Set. For example: `{"sub": "user@example.com", "iat": 313435}` If the JWT Claims Set contains an expiration time (`exp`) claim, it must be an integer timestamp that is not in the past and no more than 12 hours in the future. If the JWT Claims Set does not contain an expiration time (`exp`) claim, this claim is added automatically, with a timestamp that is 1 hour in the future.
}
x__xgafv: string, V1 error format.
diff --git a/docs/dyn/ideahub_v1alpha.ideas.html b/docs/dyn/ideahub_v1alpha.ideas.html
index a9b61b3c928..9fb75166d4f 100644
--- a/docs/dyn/ideahub_v1alpha.ideas.html
+++ b/docs/dyn/ideahub_v1alpha.ideas.html
@@ -122,6 +122,7 @@ Method Details
{ # Represents a Topic umbrella for a list of questions that a Creator may want to respond to.
"displayName": "A String", # String displayed to the creator indicating the name of the Topic.
"mid": "A String", # The mID of the topic.
+ "name": "A String", # Unique identifier for the topic. Format: topics/{topic}
},
],
},
diff --git a/docs/dyn/ideahub_v1alpha.platforms.properties.ideaActivities.html b/docs/dyn/ideahub_v1alpha.platforms.properties.ideaActivities.html
index 89536e0ae1d..da3c7bd8a55 100644
--- a/docs/dyn/ideahub_v1alpha.platforms.properties.ideaActivities.html
+++ b/docs/dyn/ideahub_v1alpha.platforms.properties.ideaActivities.html
@@ -96,11 +96,11 @@ Method Details
The object takes the form of:
{ # An idea activity entry.
- "ideaIds": [ # The Idea IDs for this entry.
+ "ideas": [ # The Idea IDs for this entry.
"A String",
],
"name": "A String", # Unique identifier for the idea activity. Format: platforms/{platform}/properties/{property}/ideaActivities/{idea_activity}
- "topicIds": [ # The Topic IDs for this entry.
+ "topics": [ # The Topic IDs for this entry.
"A String",
],
"type": "A String", # The type of activity performed.
@@ -116,11 +116,11 @@ Method Details
An object of the form:
{ # An idea activity entry.
- "ideaIds": [ # The Idea IDs for this entry.
+ "ideas": [ # The Idea IDs for this entry.
"A String",
],
"name": "A String", # Unique identifier for the idea activity. Format: platforms/{platform}/properties/{property}/ideaActivities/{idea_activity}
- "topicIds": [ # The Topic IDs for this entry.
+ "topics": [ # The Topic IDs for this entry.
"A String",
],
"type": "A String", # The type of activity performed.
diff --git a/docs/dyn/ideahub_v1alpha.platforms.properties.ideas.html b/docs/dyn/ideahub_v1alpha.platforms.properties.ideas.html
index 5f1b4badc69..fc92dcaf9ba 100644
--- a/docs/dyn/ideahub_v1alpha.platforms.properties.ideas.html
+++ b/docs/dyn/ideahub_v1alpha.platforms.properties.ideas.html
@@ -122,6 +122,7 @@ Method Details
{ # Represents a Topic umbrella for a list of questions that a Creator may want to respond to.
"displayName": "A String", # String displayed to the creator indicating the name of the Topic.
"mid": "A String", # The mID of the topic.
+ "name": "A String", # Unique identifier for the topic. Format: topics/{topic}
},
],
},
diff --git a/docs/dyn/index.md b/docs/dyn/index.md
index f951e828013..05280ec875a 100644
--- a/docs/dyn/index.md
+++ b/docs/dyn/index.md
@@ -288,6 +288,10 @@
* [v1](http://googleapis.github.io/google-api-python-client/docs/dyn/compute_v1.html)
+## contactcenterinsights
+* [v1](http://googleapis.github.io/google-api-python-client/docs/dyn/contactcenterinsights_v1.html)
+
+
## container
* [v1](http://googleapis.github.io/google-api-python-client/docs/dyn/container_v1.html)
* [v1beta1](http://googleapis.github.io/google-api-python-client/docs/dyn/container_v1beta1.html)
@@ -666,6 +670,11 @@
* [v1beta1](http://googleapis.github.io/google-api-python-client/docs/dyn/networksecurity_v1beta1.html)
+## networkservices
+* [v1](http://googleapis.github.io/google-api-python-client/docs/dyn/networkservices_v1.html)
+* [v1beta1](http://googleapis.github.io/google-api-python-client/docs/dyn/networkservices_v1beta1.html)
+
+
## notebooks
* [v1](http://googleapis.github.io/google-api-python-client/docs/dyn/notebooks_v1.html)
@@ -715,6 +724,11 @@
* [v1](http://googleapis.github.io/google-api-python-client/docs/dyn/playcustomapp_v1.html)
+## policyanalyzer
+* [v1](http://googleapis.github.io/google-api-python-client/docs/dyn/policyanalyzer_v1.html)
+* [v1beta1](http://googleapis.github.io/google-api-python-client/docs/dyn/policyanalyzer_v1beta1.html)
+
+
## policysimulator
* [v1](http://googleapis.github.io/google-api-python-client/docs/dyn/policysimulator_v1.html)
* [v1beta1](http://googleapis.github.io/google-api-python-client/docs/dyn/policysimulator_v1beta1.html)
@@ -788,7 +802,6 @@
## run
* [v1](http://googleapis.github.io/google-api-python-client/docs/dyn/run_v1.html)
* [v1alpha1](http://googleapis.github.io/google-api-python-client/docs/dyn/run_v1alpha1.html)
-* [v1beta1](http://googleapis.github.io/google-api-python-client/docs/dyn/run_v1beta1.html)
## runtimeconfig
diff --git a/docs/dyn/jobs_v3.projects.jobs.html b/docs/dyn/jobs_v3.projects.jobs.html
index c817cb44d82..7828e366b39 100644
--- a/docs/dyn/jobs_v3.projects.jobs.html
+++ b/docs/dyn/jobs_v3.projects.jobs.html
@@ -629,7 +629,7 @@ Method Details
Args:
parent: string, Required. The resource name of the project under which the job is created. The format is "projects/{project_id}", for example, "projects/api-test-project". (required)
- filter: string, Required. The filter string specifies the jobs to be enumerated. Supported operator: =, AND The fields eligible for filtering are: * `companyName` (Required) * `requisitionId` (Optional) Sample Query: * companyName = "projects/api-test-project/companies/123" * companyName = "projects/api-test-project/companies/123" AND requisitionId = "req-1"
+ filter: string, Required. The filter string specifies the jobs to be enumerated. Supported operator: =, AND The fields eligible for filtering are: * `companyName` * `requisitionId` * `status` Available values: OPEN, EXPIRED, ALL. Defaults to OPEN if no value is specified. At least one of `companyName` and `requisitionId` must present or an INVALID_ARGUMENT error is thrown. Sample Query: * companyName = "projects/api-test-project/companies/123" * companyName = "projects/api-test-project/companies/123" AND requisitionId = "req-1" * companyName = "projects/api-test-project/companies/123" AND status = "EXPIRED" * requisitionId = "req-1" * requisitionId = "req-1" AND status = "EXPIRED"
jobView: string, Optional. The desired job attributes returned for jobs in the search response. Defaults to JobView.JOB_VIEW_FULL if no value is specified.
Allowed values
JOB_VIEW_UNSPECIFIED - Default value.
diff --git a/docs/dyn/jobs_v3p1beta1.projects.jobs.html b/docs/dyn/jobs_v3p1beta1.projects.jobs.html
index 9face680787..acb681d31a9 100644
--- a/docs/dyn/jobs_v3p1beta1.projects.jobs.html
+++ b/docs/dyn/jobs_v3p1beta1.projects.jobs.html
@@ -629,7 +629,7 @@ Method Details
Args:
parent: string, Required. The resource name of the project under which the job is created. The format is "projects/{project_id}", for example, "projects/api-test-project". (required)
- filter: string, Required. The filter string specifies the jobs to be enumerated. Supported operator: =, AND The fields eligible for filtering are: * `companyName` (Required) * `requisitionId` (Optional) Sample Query: * companyName = "projects/api-test-project/companies/123" * companyName = "projects/api-test-project/companies/123" AND requisitionId = "req-1"
+ filter: string, Required. The filter string specifies the jobs to be enumerated. Supported operator: =, AND The fields eligible for filtering are: * `companyName` * `requisitionId` * `status` Available values: OPEN, EXPIRED, ALL. Defaults to OPEN if no value is specified. At least one of `companyName` and `requisitionId` must present or an INVALID_ARGUMENT error is thrown. Sample Query: * companyName = "projects/api-test-project/companies/123" * companyName = "projects/api-test-project/companies/123" AND requisitionId = "req-1" * companyName = "projects/api-test-project/companies/123" AND status = "EXPIRED" * requisitionId = "req-1" * requisitionId = "req-1" AND status = "EXPIRED"
jobView: string, Optional. The desired job attributes returned for jobs in the search response. Defaults to JobView.JOB_VIEW_FULL if no value is specified.
Allowed values
JOB_VIEW_UNSPECIFIED - Default value.
diff --git a/docs/dyn/jobs_v4.projects.tenants.jobs.html b/docs/dyn/jobs_v4.projects.tenants.jobs.html
index e32e5a9a8da..7337c8b25ee 100644
--- a/docs/dyn/jobs_v4.projects.tenants.jobs.html
+++ b/docs/dyn/jobs_v4.projects.tenants.jobs.html
@@ -1028,7 +1028,7 @@ Method Details
Args:
parent: string, Required. The resource name of the tenant under which the job is created. The format is "projects/{project_id}/tenants/{tenant_id}". For example, "projects/foo/tenants/bar". (required)
- filter: string, Required. The filter string specifies the jobs to be enumerated. Supported operator: =, AND The fields eligible for filtering are: * `companyName` (Required) * `requisitionId` * `status` Available values: OPEN, EXPIRED, ALL. Defaults to OPEN if no value is specified. Sample Query: * companyName = "projects/foo/tenants/bar/companies/baz" * companyName = "projects/foo/tenants/bar/companies/baz" AND requisitionId = "req-1" * companyName = "projects/foo/tenants/bar/companies/baz" AND status = "EXPIRED"
+ filter: string, Required. The filter string specifies the jobs to be enumerated. Supported operator: =, AND The fields eligible for filtering are: * `companyName` * `requisitionId` * `status` Available values: OPEN, EXPIRED, ALL. Defaults to OPEN if no value is specified. At least one of `companyName` and `requisitionId` must present or an INVALID_ARGUMENT error is thrown. Sample Query: * companyName = "projects/foo/tenants/bar/companies/baz" * companyName = "projects/foo/tenants/bar/companies/baz" AND requisitionId = "req-1" * companyName = "projects/foo/tenants/bar/companies/baz" AND status = "EXPIRED" * requisitionId = "req-1" * requisitionId = "req-1" AND status = "EXPIRED"
jobView: string, The desired job attributes returned for jobs in the search response. Defaults to JobView.JOB_VIEW_FULL if no value is specified.
Allowed values
JOB_VIEW_UNSPECIFIED - Default value.
diff --git a/docs/dyn/logging_v2.billingAccounts.buckets.html b/docs/dyn/logging_v2.billingAccounts.buckets.html
index 2a04fa476e4..195d0cc8401 100644
--- a/docs/dyn/logging_v2.billingAccounts.buckets.html
+++ b/docs/dyn/logging_v2.billingAccounts.buckets.html
@@ -84,7 +84,7 @@ Instance Methods
Close httplib2 connections.
-Gets a bucket.
+Gets a log bucket.
Method Details
close()
@@ -93,10 +93,10 @@ Method Details
get(name, x__xgafv=None)
- Gets a bucket.
+ Gets a log bucket.
Args:
- name: string, Required. The resource name of the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The resource name of the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -105,13 +105,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
diff --git a/docs/dyn/logging_v2.billingAccounts.buckets.views.html b/docs/dyn/logging_v2.billingAccounts.buckets.views.html
index 804a0aaa9cb..31b19740ca0 100644
--- a/docs/dyn/logging_v2.billingAccounts.buckets.views.html
+++ b/docs/dyn/logging_v2.billingAccounts.buckets.views.html
@@ -79,7 +79,7 @@ Instance Methods
Close httplib2 connections.
-Gets a view.
+Gets a view on a log bucket..
Method Details
close()
@@ -88,10 +88,10 @@ Method Details
get(name, x__xgafv=None)
- Gets a view.
+ Gets a view on a log bucket..
Args:
- name: string, Required. The resource name of the policy: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The resource name of the policy: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -100,11 +100,11 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
diff --git a/docs/dyn/logging_v2.billingAccounts.exclusions.html b/docs/dyn/logging_v2.billingAccounts.exclusions.html
index bcda68858d5..85c8c7a0523 100644
--- a/docs/dyn/logging_v2.billingAccounts.exclusions.html
+++ b/docs/dyn/logging_v2.billingAccounts.exclusions.html
@@ -106,15 +106,15 @@ Method Details
Creates a new exclusion in a specified parent resource. Only log entries belonging to that resource can be excluded. You can have up to 10 exclusions in a resource.
Args:
- parent: string, Required. The parent resource in which to create the exclusion: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" Examples: "projects/my-logging-project", "organizations/123456789". (required)
+ parent: string, Required. The parent resource in which to create the exclusion: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" For examples:"projects/my-logging-project" "organizations/123456789" (required)
body: object, The request body.
The object takes the form of:
-{ # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+{ # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -127,11 +127,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -142,7 +142,7 @@ Method Details
Deletes an exclusion.
Args:
- name: string, Required. The resource name of an existing exclusion to delete: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of an existing exclusion to delete: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -160,7 +160,7 @@ Method Details
Gets the description of an exclusion.
Args:
- name: string, Required. The resource name of an existing exclusion: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of an existing exclusion: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -169,11 +169,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -197,11 +197,11 @@ Method Details
{ # Result returned from ListExclusions.
"exclusions": [ # A list of exclusions.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
@@ -229,15 +229,15 @@ Method Details
Changes one or more properties of an existing exclusion.
Args:
- name: string, Required. The resource name of the exclusion to update: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of the exclusion to update: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
body: object, The request body.
The object takes the form of:
-{ # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+{ # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -251,11 +251,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
diff --git a/docs/dyn/logging_v2.billingAccounts.locations.buckets.html b/docs/dyn/logging_v2.billingAccounts.locations.buckets.html
index 6beadf0f84f..0840e7bb2ef 100644
--- a/docs/dyn/logging_v2.billingAccounts.locations.buckets.html
+++ b/docs/dyn/logging_v2.billingAccounts.locations.buckets.html
@@ -84,22 +84,22 @@ Instance Methods
Close httplib2 connections.
create(parent, body=None, bucketId=None, x__xgafv=None)
-Creates a bucket that can be used to store log entries. Once a bucket has been created, the region cannot be changed.
+Creates a log bucket that can be used to store log entries. After a bucket has been created, the bucket's location cannot be changed.
-Deletes a bucket. Moves the bucket to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all logs in the bucket will be permanently deleted.
+Deletes a log bucket.Changes the bucket's lifecycle_state to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all log entries in the bucket will be permanently deleted.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists buckets.
+Lists log buckets.
list_next(previous_request, previous_response)
Retrieves the next page of results.
patch(name, body=None, updateMask=None, x__xgafv=None)
-Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created.
+Updates a log bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a lifecycle_state of DELETE_REQUESTED, then FAILED_PRECONDITION will be returned.After a bucket has been created, the bucket's location cannot be changed.
undelete(name, body=None, x__xgafv=None)
-Undeletes a bucket. A bucket that has been deleted may be undeleted within the grace period of 7 days.
+Undeletes a log bucket. A bucket that has been deleted can be undeleted within the grace period of 7 days.
Method Details
close()
@@ -108,20 +108,20 @@ Method Details
create(parent, body=None, bucketId=None, x__xgafv=None)
- Creates a bucket that can be used to store log entries. Once a bucket has been created, the region cannot be changed.
+ Creates a log bucket that can be used to store log entries. After a bucket has been created, the bucket's location cannot be changed.
Args:
- parent: string, Required. The resource in which to create the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" Example: "projects/my-logging-project/locations/global" (required)
+ parent: string, Required. The resource in which to create the log bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" For example:"projects/my-project/locations/global" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a repository of logs.
+{ # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -137,13 +137,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -153,10 +153,10 @@ Method Details
delete(name, x__xgafv=None)
- Deletes a bucket. Moves the bucket to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all logs in the bucket will be permanently deleted.
+ Deletes a log bucket.Changes the bucket's lifecycle_state to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all log entries in the bucket will be permanently deleted.
Args:
- name: string, Required. The full resource name of the bucket to delete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The full resource name of the bucket to delete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -171,7 +171,7 @@ Method Details
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists buckets.
+ Lists log buckets.
Args:
parent: string, Required. The parent resource whose buckets are to be listed: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]" Note: The locations portion of the resource must be specified, but supplying the character - in place of LOCATION_ID will return all buckets. (required)
@@ -187,13 +187,13 @@ Method Details
{ # The response from ListBuckets.
"buckets": [ # A list of buckets.
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -220,27 +220,27 @@ Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
- Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created.
+ Updates a log bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a lifecycle_state of DELETE_REQUESTED, then FAILED_PRECONDITION will be returned.After a bucket has been created, the bucket's location cannot be changed.
Args:
- name: string, Required. The full resource name of the bucket to update. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". Also requires permission "resourcemanager.projects.updateLiens" to set the locked property (required)
+ name: string, Required. The full resource name of the bucket to update. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a repository of logs.
+{ # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
"updateTime": "A String", # Output only. The last update timestamp of the bucket.
}
- updateMask: string, Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=retention_days.
+ updateMask: string, Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see: https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=retention_days
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -249,13 +249,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -265,10 +265,10 @@ Method Details
undelete(name, body=None, x__xgafv=None)
- Undeletes a bucket. A bucket that has been deleted may be undeleted within the grace period of 7 days.
+ Undeletes a log bucket. A bucket that has been deleted can be undeleted within the grace period of 7 days.
Args:
- name: string, Required. The full resource name of the bucket to undelete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The full resource name of the bucket to undelete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
diff --git a/docs/dyn/logging_v2.billingAccounts.locations.buckets.views.html b/docs/dyn/logging_v2.billingAccounts.locations.buckets.views.html
index 1ec1edc7fff..e05c8380919 100644
--- a/docs/dyn/logging_v2.billingAccounts.locations.buckets.views.html
+++ b/docs/dyn/logging_v2.billingAccounts.locations.buckets.views.html
@@ -79,19 +79,19 @@ Instance Methods
Close httplib2 connections.
create(parent, body=None, viewId=None, x__xgafv=None)
-Creates a view over logs in a bucket. A bucket may contain a maximum of 50 views.
+Creates a view over log entries in a log bucket. A bucket may contain a maximum of 50 views.
-Deletes a view from a bucket.
+Deletes a view on a log bucket.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists views on a bucket.
+Lists views on a log bucket.
list_next(previous_request, previous_response)
Retrieves the next page of results.
patch(name, body=None, updateMask=None, x__xgafv=None)
-Updates a view. This method replaces the following fields in the existing view with values from the new view: filter.
+Updates a view on a log bucket. This method replaces the following fields in the existing view with values from the new view: filter.
Method Details
close()
@@ -100,18 +100,18 @@ Method Details
create(parent, body=None, viewId=None, x__xgafv=None)
- Creates a view over logs in a bucket. A bucket may contain a maximum of 50 views.
+ Creates a view over log entries in a log bucket. A bucket may contain a maximum of 50 views.
Args:
- parent: string, Required. The bucket in which to create the view "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-logging-project/locations/my-location/buckets/my-bucket" (required)
+ parent: string, Required. The bucket in which to create the view `"projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]"` For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a view over logs in a bucket.
+{ # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
@@ -124,21 +124,21 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
delete(name, x__xgafv=None)
- Deletes a view from a bucket.
+ Deletes a view on a log bucket.
Args:
- name: string, Required. The full resource name of the view to delete: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The full resource name of the view to delete: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -153,11 +153,11 @@ Method Details
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists views on a bucket.
+ Lists views on a log bucket.
Args:
parent: string, Required. The bucket whose views are to be listed: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" (required)
- pageSize: integer, Optional. The maximum number of results to return from this request. Non-positive values are ignored. The presence of nextPageToken in the response indicates that more results might be available.
+ pageSize: integer, Optional. The maximum number of results to return from this request.Non-positive values are ignored. The presence of nextPageToken in the response indicates that more results might be available.
pageToken: string, Optional. If present, then retrieve the next batch of results from the preceding call to this method. pageToken must be the value of nextPageToken from the previous response. The values of other method parameters should be identical to those in the previous call.
x__xgafv: string, V1 error format.
Allowed values
@@ -170,11 +170,11 @@ Method Details
{ # The response from ListViews.
"nextPageToken": "A String", # If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken.
"views": [ # A list of views.
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
},
],
@@ -197,22 +197,22 @@ Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
- Updates a view. This method replaces the following fields in the existing view with values from the new view: filter.
+ Updates a view on a log bucket. This method replaces the following fields in the existing view with values from the new view: filter.
Args:
- name: string, Required. The full resource name of the view to update "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The full resource name of the view to update "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a view over logs in a bucket.
+{ # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
- updateMask: string, Optional. Field mask that specifies the fields in view that need an update. A field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in view that need an update. A field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -221,11 +221,11 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
diff --git a/docs/dyn/logging_v2.billingAccounts.sinks.html b/docs/dyn/logging_v2.billingAccounts.sinks.html
index 3897a6808b9..738158905e2 100644
--- a/docs/dyn/logging_v2.billingAccounts.sinks.html
+++ b/docs/dyn/logging_v2.billingAccounts.sinks.html
@@ -109,38 +109,38 @@ Method Details
Creates a sink that exports specified log entries to a destination. The export of newly-ingested log entries begins immediately, unless the sink's writer_identity is not permitted to write to the destination. A sink can export log entries only from the resource owning the sink.
Args:
- parent: string, Required. The resource in which to create the sink: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" Examples: "projects/my-logging-project", "organizations/123456789". (required)
+ parent: string, Required. The resource in which to create the sink: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" For examples:"projects/my-project" "organizations/123456789" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
- uniqueWriterIdentity: boolean, Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.
+ uniqueWriterIdentity: boolean, Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Cloud Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -149,31 +149,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -182,7 +182,7 @@ Method Details
Deletes a sink. If the sink has a unique writer_identity, then that service account is also deleted.
Args:
- sinkName: string, Required. The full resource name of the sink to delete, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to delete, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -200,7 +200,7 @@ Method Details
Gets a sink.
Args:
- sinkName: string, Required. The resource name of the sink: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The resource name of the sink: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -209,31 +209,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -256,31 +256,31 @@ Method Details
{ # Result returned from ListSinks.
"nextPageToken": "A String", # If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken.
"sinks": [ # A list of sinks.
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
},
],
}
@@ -305,39 +305,39 @@ Method Details
Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.
Args:
- sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
uniqueWriterIdentity: boolean, Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field: If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity. If the old value is false and the new value is true, then writer_identity is changed to a unique service account. It is an error if the old value is true and the new value is set to false or defaulted to false.
- updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes:destination,filter,includeChildrenAt some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -346,31 +346,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -379,39 +379,39 @@ Method Details
Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.
Args:
- sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
uniqueWriterIdentity: boolean, Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field: If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity. If the old value is false and the new value is true, then writer_identity is changed to a unique service account. It is an error if the old value is true and the new value is set to false or defaulted to false.
- updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes:destination,filter,includeChildrenAt some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -420,31 +420,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
diff --git a/docs/dyn/logging_v2.entries.html b/docs/dyn/logging_v2.entries.html
index fa2d57be302..7a4df635c21 100644
--- a/docs/dyn/logging_v2.entries.html
+++ b/docs/dyn/logging_v2.entries.html
@@ -79,7 +79,7 @@ Instance Methods
Close httplib2 connections.
copy(body=None, x__xgafv=None)
-Copies a set of log entries from a logging bucket to a Cloud Storage bucket.
+Copies a set of log entries from a log bucket to a Cloud Storage bucket.
list(body=None, x__xgafv=None)
Lists log entries. Use this method to retrieve log entries that originated from a project/folder/organization/billing account. For ways to export log entries, see Exporting Logs (https://cloud.google.com/logging/docs/export).
@@ -100,16 +100,16 @@ Method Details
copy(body=None, x__xgafv=None)
- Copies a set of log entries from a logging bucket to a Cloud Storage bucket.
+ Copies a set of log entries from a log bucket to a Cloud Storage bucket.
Args:
body: object, The request body.
The object takes the form of:
{ # The parameters to CopyLogEntries.
- "destination": "A String", # Required. Destination to which to copy logs.
+ "destination": "A String", # Required. Destination to which to copy log entries.
"filter": "A String", # Optional. A filter specifying which log entries to copy. The filter must be no more than 20k characters. An empty filter matches all log entries.
- "name": "A String", # Required. Bucket from which to copy logs. e.g. "projects/my-project/locations/my-location/buckets/my-source-bucket
+ "name": "A String", # Required. Log bucket from which to copy log entries.For example:"projects/my-project/locations/global/buckets/my-source-bucket"
}
x__xgafv: string, V1 error format.
diff --git a/docs/dyn/logging_v2.exclusions.html b/docs/dyn/logging_v2.exclusions.html
index 044f1db8377..f42a40549eb 100644
--- a/docs/dyn/logging_v2.exclusions.html
+++ b/docs/dyn/logging_v2.exclusions.html
@@ -106,15 +106,15 @@ Method Details
Creates a new exclusion in a specified parent resource. Only log entries belonging to that resource can be excluded. You can have up to 10 exclusions in a resource.
Args:
- parent: string, Required. The parent resource in which to create the exclusion: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" Examples: "projects/my-logging-project", "organizations/123456789". (required)
+ parent: string, Required. The parent resource in which to create the exclusion: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" For examples:"projects/my-logging-project" "organizations/123456789" (required)
body: object, The request body.
The object takes the form of:
-{ # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+{ # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -127,11 +127,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -142,7 +142,7 @@ Method Details
Deletes an exclusion.
Args:
- name: string, Required. The resource name of an existing exclusion to delete: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of an existing exclusion to delete: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -160,7 +160,7 @@ Method Details
Gets the description of an exclusion.
Args:
- name: string, Required. The resource name of an existing exclusion: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of an existing exclusion: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -169,11 +169,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -197,11 +197,11 @@ Method Details
{ # Result returned from ListExclusions.
"exclusions": [ # A list of exclusions.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
@@ -229,15 +229,15 @@ Method Details
Changes one or more properties of an existing exclusion.
Args:
- name: string, Required. The resource name of the exclusion to update: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of the exclusion to update: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
body: object, The request body.
The object takes the form of:
-{ # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+{ # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -251,11 +251,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
diff --git a/docs/dyn/logging_v2.folders.exclusions.html b/docs/dyn/logging_v2.folders.exclusions.html
index 7c18f882ffd..ba514fda40d 100644
--- a/docs/dyn/logging_v2.folders.exclusions.html
+++ b/docs/dyn/logging_v2.folders.exclusions.html
@@ -106,15 +106,15 @@ Method Details
Creates a new exclusion in a specified parent resource. Only log entries belonging to that resource can be excluded. You can have up to 10 exclusions in a resource.
Args:
- parent: string, Required. The parent resource in which to create the exclusion: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" Examples: "projects/my-logging-project", "organizations/123456789". (required)
+ parent: string, Required. The parent resource in which to create the exclusion: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" For examples:"projects/my-logging-project" "organizations/123456789" (required)
body: object, The request body.
The object takes the form of:
-{ # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+{ # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -127,11 +127,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -142,7 +142,7 @@ Method Details
Deletes an exclusion.
Args:
- name: string, Required. The resource name of an existing exclusion to delete: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of an existing exclusion to delete: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -160,7 +160,7 @@ Method Details
Gets the description of an exclusion.
Args:
- name: string, Required. The resource name of an existing exclusion: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of an existing exclusion: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -169,11 +169,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -197,11 +197,11 @@ Method Details
{ # Result returned from ListExclusions.
"exclusions": [ # A list of exclusions.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
@@ -229,15 +229,15 @@ Method Details
Changes one or more properties of an existing exclusion.
Args:
- name: string, Required. The resource name of the exclusion to update: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of the exclusion to update: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
body: object, The request body.
The object takes the form of:
-{ # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+{ # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -251,11 +251,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
diff --git a/docs/dyn/logging_v2.folders.locations.buckets.html b/docs/dyn/logging_v2.folders.locations.buckets.html
index 135108a9419..3884f615153 100644
--- a/docs/dyn/logging_v2.folders.locations.buckets.html
+++ b/docs/dyn/logging_v2.folders.locations.buckets.html
@@ -84,25 +84,25 @@ Instance Methods
Close httplib2 connections.
create(parent, body=None, bucketId=None, x__xgafv=None)
-Creates a bucket that can be used to store log entries. Once a bucket has been created, the region cannot be changed.
+Creates a log bucket that can be used to store log entries. After a bucket has been created, the bucket's location cannot be changed.
-Deletes a bucket. Moves the bucket to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all logs in the bucket will be permanently deleted.
+Deletes a log bucket.Changes the bucket's lifecycle_state to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all log entries in the bucket will be permanently deleted.
-Gets a bucket.
+Gets a log bucket.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists buckets.
+Lists log buckets.
list_next(previous_request, previous_response)
Retrieves the next page of results.
patch(name, body=None, updateMask=None, x__xgafv=None)
-Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created.
+Updates a log bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a lifecycle_state of DELETE_REQUESTED, then FAILED_PRECONDITION will be returned.After a bucket has been created, the bucket's location cannot be changed.
undelete(name, body=None, x__xgafv=None)
-Undeletes a bucket. A bucket that has been deleted may be undeleted within the grace period of 7 days.
+Undeletes a log bucket. A bucket that has been deleted can be undeleted within the grace period of 7 days.
Method Details
close()
@@ -111,20 +111,20 @@ Method Details
create(parent, body=None, bucketId=None, x__xgafv=None)
- Creates a bucket that can be used to store log entries. Once a bucket has been created, the region cannot be changed.
+ Creates a log bucket that can be used to store log entries. After a bucket has been created, the bucket's location cannot be changed.
Args:
- parent: string, Required. The resource in which to create the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" Example: "projects/my-logging-project/locations/global" (required)
+ parent: string, Required. The resource in which to create the log bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" For example:"projects/my-project/locations/global" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a repository of logs.
+{ # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -140,13 +140,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -156,10 +156,10 @@ Method Details
delete(name, x__xgafv=None)
- Deletes a bucket. Moves the bucket to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all logs in the bucket will be permanently deleted.
+ Deletes a log bucket.Changes the bucket's lifecycle_state to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all log entries in the bucket will be permanently deleted.
Args:
- name: string, Required. The full resource name of the bucket to delete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The full resource name of the bucket to delete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -174,10 +174,10 @@ Method Details
get(name, x__xgafv=None)
- Gets a bucket.
+ Gets a log bucket.
Args:
- name: string, Required. The resource name of the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The resource name of the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -186,13 +186,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -202,7 +202,7 @@ Method Details
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists buckets.
+ Lists log buckets.
Args:
parent: string, Required. The parent resource whose buckets are to be listed: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]" Note: The locations portion of the resource must be specified, but supplying the character - in place of LOCATION_ID will return all buckets. (required)
@@ -218,13 +218,13 @@ Method Details
{ # The response from ListBuckets.
"buckets": [ # A list of buckets.
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -251,27 +251,27 @@ Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
- Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created.
+ Updates a log bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a lifecycle_state of DELETE_REQUESTED, then FAILED_PRECONDITION will be returned.After a bucket has been created, the bucket's location cannot be changed.
Args:
- name: string, Required. The full resource name of the bucket to update. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". Also requires permission "resourcemanager.projects.updateLiens" to set the locked property (required)
+ name: string, Required. The full resource name of the bucket to update. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a repository of logs.
+{ # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
"updateTime": "A String", # Output only. The last update timestamp of the bucket.
}
- updateMask: string, Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=retention_days.
+ updateMask: string, Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see: https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=retention_days
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -280,13 +280,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -296,10 +296,10 @@ Method Details
undelete(name, body=None, x__xgafv=None)
- Undeletes a bucket. A bucket that has been deleted may be undeleted within the grace period of 7 days.
+ Undeletes a log bucket. A bucket that has been deleted can be undeleted within the grace period of 7 days.
Args:
- name: string, Required. The full resource name of the bucket to undelete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The full resource name of the bucket to undelete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
diff --git a/docs/dyn/logging_v2.folders.locations.buckets.views.html b/docs/dyn/logging_v2.folders.locations.buckets.views.html
index 263918332f6..28f4c421384 100644
--- a/docs/dyn/logging_v2.folders.locations.buckets.views.html
+++ b/docs/dyn/logging_v2.folders.locations.buckets.views.html
@@ -79,22 +79,22 @@ Instance Methods
Close httplib2 connections.
create(parent, body=None, viewId=None, x__xgafv=None)
-Creates a view over logs in a bucket. A bucket may contain a maximum of 50 views.
+Creates a view over log entries in a log bucket. A bucket may contain a maximum of 50 views.
-Deletes a view from a bucket.
+Deletes a view on a log bucket.
-Gets a view.
+Gets a view on a log bucket..
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists views on a bucket.
+Lists views on a log bucket.
list_next(previous_request, previous_response)
Retrieves the next page of results.
patch(name, body=None, updateMask=None, x__xgafv=None)
-Updates a view. This method replaces the following fields in the existing view with values from the new view: filter.
+Updates a view on a log bucket. This method replaces the following fields in the existing view with values from the new view: filter.
Method Details
close()
@@ -103,18 +103,18 @@ Method Details
create(parent, body=None, viewId=None, x__xgafv=None)
- Creates a view over logs in a bucket. A bucket may contain a maximum of 50 views.
+ Creates a view over log entries in a log bucket. A bucket may contain a maximum of 50 views.
Args:
- parent: string, Required. The bucket in which to create the view "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-logging-project/locations/my-location/buckets/my-bucket" (required)
+ parent: string, Required. The bucket in which to create the view `"projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]"` For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a view over logs in a bucket.
+{ # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
@@ -127,21 +127,21 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
delete(name, x__xgafv=None)
- Deletes a view from a bucket.
+ Deletes a view on a log bucket.
Args:
- name: string, Required. The full resource name of the view to delete: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The full resource name of the view to delete: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -156,10 +156,10 @@ Method Details
get(name, x__xgafv=None)
- Gets a view.
+ Gets a view on a log bucket..
Args:
- name: string, Required. The resource name of the policy: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The resource name of the policy: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -168,22 +168,22 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists views on a bucket.
+ Lists views on a log bucket.
Args:
parent: string, Required. The bucket whose views are to be listed: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" (required)
- pageSize: integer, Optional. The maximum number of results to return from this request. Non-positive values are ignored. The presence of nextPageToken in the response indicates that more results might be available.
+ pageSize: integer, Optional. The maximum number of results to return from this request.Non-positive values are ignored. The presence of nextPageToken in the response indicates that more results might be available.
pageToken: string, Optional. If present, then retrieve the next batch of results from the preceding call to this method. pageToken must be the value of nextPageToken from the previous response. The values of other method parameters should be identical to those in the previous call.
x__xgafv: string, V1 error format.
Allowed values
@@ -196,11 +196,11 @@ Method Details
{ # The response from ListViews.
"nextPageToken": "A String", # If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken.
"views": [ # A list of views.
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
},
],
@@ -223,22 +223,22 @@ Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
- Updates a view. This method replaces the following fields in the existing view with values from the new view: filter.
+ Updates a view on a log bucket. This method replaces the following fields in the existing view with values from the new view: filter.
Args:
- name: string, Required. The full resource name of the view to update "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The full resource name of the view to update "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a view over logs in a bucket.
+{ # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
- updateMask: string, Optional. Field mask that specifies the fields in view that need an update. A field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in view that need an update. A field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -247,11 +247,11 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
diff --git a/docs/dyn/logging_v2.folders.sinks.html b/docs/dyn/logging_v2.folders.sinks.html
index 2de320420b9..2d1a69b4769 100644
--- a/docs/dyn/logging_v2.folders.sinks.html
+++ b/docs/dyn/logging_v2.folders.sinks.html
@@ -109,38 +109,38 @@ Method Details
Creates a sink that exports specified log entries to a destination. The export of newly-ingested log entries begins immediately, unless the sink's writer_identity is not permitted to write to the destination. A sink can export log entries only from the resource owning the sink.
Args:
- parent: string, Required. The resource in which to create the sink: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" Examples: "projects/my-logging-project", "organizations/123456789". (required)
+ parent: string, Required. The resource in which to create the sink: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" For examples:"projects/my-project" "organizations/123456789" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
- uniqueWriterIdentity: boolean, Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.
+ uniqueWriterIdentity: boolean, Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Cloud Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -149,31 +149,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -182,7 +182,7 @@ Method Details
Deletes a sink. If the sink has a unique writer_identity, then that service account is also deleted.
Args:
- sinkName: string, Required. The full resource name of the sink to delete, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to delete, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -200,7 +200,7 @@ Method Details
Gets a sink.
Args:
- sinkName: string, Required. The resource name of the sink: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The resource name of the sink: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -209,31 +209,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -256,31 +256,31 @@ Method Details
{ # Result returned from ListSinks.
"nextPageToken": "A String", # If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken.
"sinks": [ # A list of sinks.
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
},
],
}
@@ -305,39 +305,39 @@ Method Details
Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.
Args:
- sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
uniqueWriterIdentity: boolean, Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field: If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity. If the old value is false and the new value is true, then writer_identity is changed to a unique service account. It is an error if the old value is true and the new value is set to false or defaulted to false.
- updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes:destination,filter,includeChildrenAt some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -346,31 +346,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -379,39 +379,39 @@ Method Details
Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.
Args:
- sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
uniqueWriterIdentity: boolean, Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field: If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity. If the old value is false and the new value is true, then writer_identity is changed to a unique service account. It is an error if the old value is true and the new value is set to false or defaulted to false.
- updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes:destination,filter,includeChildrenAt some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -420,31 +420,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
diff --git a/docs/dyn/logging_v2.locations.buckets.html b/docs/dyn/logging_v2.locations.buckets.html
index 1e501d3d128..6d950fa52c5 100644
--- a/docs/dyn/logging_v2.locations.buckets.html
+++ b/docs/dyn/logging_v2.locations.buckets.html
@@ -84,25 +84,25 @@ Instance Methods
Close httplib2 connections.
create(parent, body=None, bucketId=None, x__xgafv=None)
-Creates a bucket that can be used to store log entries. Once a bucket has been created, the region cannot be changed.
+Creates a log bucket that can be used to store log entries. After a bucket has been created, the bucket's location cannot be changed.
-Deletes a bucket. Moves the bucket to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all logs in the bucket will be permanently deleted.
+Deletes a log bucket.Changes the bucket's lifecycle_state to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all log entries in the bucket will be permanently deleted.
-Gets a bucket.
+Gets a log bucket.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists buckets.
+Lists log buckets.
list_next(previous_request, previous_response)
Retrieves the next page of results.
patch(name, body=None, updateMask=None, x__xgafv=None)
-Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created.
+Updates a log bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a lifecycle_state of DELETE_REQUESTED, then FAILED_PRECONDITION will be returned.After a bucket has been created, the bucket's location cannot be changed.
undelete(name, body=None, x__xgafv=None)
-Undeletes a bucket. A bucket that has been deleted may be undeleted within the grace period of 7 days.
+Undeletes a log bucket. A bucket that has been deleted can be undeleted within the grace period of 7 days.
Method Details
close()
@@ -111,20 +111,20 @@ Method Details
create(parent, body=None, bucketId=None, x__xgafv=None)
- Creates a bucket that can be used to store log entries. Once a bucket has been created, the region cannot be changed.
+ Creates a log bucket that can be used to store log entries. After a bucket has been created, the bucket's location cannot be changed.
Args:
- parent: string, Required. The resource in which to create the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" Example: "projects/my-logging-project/locations/global" (required)
+ parent: string, Required. The resource in which to create the log bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" For example:"projects/my-project/locations/global" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a repository of logs.
+{ # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -140,13 +140,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -156,10 +156,10 @@ Method Details
delete(name, x__xgafv=None)
- Deletes a bucket. Moves the bucket to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all logs in the bucket will be permanently deleted.
+ Deletes a log bucket.Changes the bucket's lifecycle_state to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all log entries in the bucket will be permanently deleted.
Args:
- name: string, Required. The full resource name of the bucket to delete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The full resource name of the bucket to delete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -174,10 +174,10 @@ Method Details
get(name, x__xgafv=None)
- Gets a bucket.
+ Gets a log bucket.
Args:
- name: string, Required. The resource name of the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The resource name of the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -186,13 +186,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -202,7 +202,7 @@ Method Details
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists buckets.
+ Lists log buckets.
Args:
parent: string, Required. The parent resource whose buckets are to be listed: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]" Note: The locations portion of the resource must be specified, but supplying the character - in place of LOCATION_ID will return all buckets. (required)
@@ -218,13 +218,13 @@ Method Details
{ # The response from ListBuckets.
"buckets": [ # A list of buckets.
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -251,27 +251,27 @@ Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
- Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created.
+ Updates a log bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a lifecycle_state of DELETE_REQUESTED, then FAILED_PRECONDITION will be returned.After a bucket has been created, the bucket's location cannot be changed.
Args:
- name: string, Required. The full resource name of the bucket to update. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". Also requires permission "resourcemanager.projects.updateLiens" to set the locked property (required)
+ name: string, Required. The full resource name of the bucket to update. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a repository of logs.
+{ # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
"updateTime": "A String", # Output only. The last update timestamp of the bucket.
}
- updateMask: string, Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=retention_days.
+ updateMask: string, Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see: https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=retention_days
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -280,13 +280,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -296,10 +296,10 @@ Method Details
undelete(name, body=None, x__xgafv=None)
- Undeletes a bucket. A bucket that has been deleted may be undeleted within the grace period of 7 days.
+ Undeletes a log bucket. A bucket that has been deleted can be undeleted within the grace period of 7 days.
Args:
- name: string, Required. The full resource name of the bucket to undelete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The full resource name of the bucket to undelete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
diff --git a/docs/dyn/logging_v2.locations.buckets.views.html b/docs/dyn/logging_v2.locations.buckets.views.html
index e77b2a91f7f..d06be986637 100644
--- a/docs/dyn/logging_v2.locations.buckets.views.html
+++ b/docs/dyn/logging_v2.locations.buckets.views.html
@@ -79,22 +79,22 @@ Instance Methods
Close httplib2 connections.
create(parent, body=None, viewId=None, x__xgafv=None)
-Creates a view over logs in a bucket. A bucket may contain a maximum of 50 views.
+Creates a view over log entries in a log bucket. A bucket may contain a maximum of 50 views.
-Deletes a view from a bucket.
+Deletes a view on a log bucket.
-Gets a view.
+Gets a view on a log bucket..
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists views on a bucket.
+Lists views on a log bucket.
list_next(previous_request, previous_response)
Retrieves the next page of results.
patch(name, body=None, updateMask=None, x__xgafv=None)
-Updates a view. This method replaces the following fields in the existing view with values from the new view: filter.
+Updates a view on a log bucket. This method replaces the following fields in the existing view with values from the new view: filter.
Method Details
close()
@@ -103,18 +103,18 @@ Method Details
create(parent, body=None, viewId=None, x__xgafv=None)
- Creates a view over logs in a bucket. A bucket may contain a maximum of 50 views.
+ Creates a view over log entries in a log bucket. A bucket may contain a maximum of 50 views.
Args:
- parent: string, Required. The bucket in which to create the view "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-logging-project/locations/my-location/buckets/my-bucket" (required)
+ parent: string, Required. The bucket in which to create the view `"projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]"` For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a view over logs in a bucket.
+{ # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
@@ -127,21 +127,21 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
delete(name, x__xgafv=None)
- Deletes a view from a bucket.
+ Deletes a view on a log bucket.
Args:
- name: string, Required. The full resource name of the view to delete: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The full resource name of the view to delete: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -156,10 +156,10 @@ Method Details
get(name, x__xgafv=None)
- Gets a view.
+ Gets a view on a log bucket..
Args:
- name: string, Required. The resource name of the policy: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The resource name of the policy: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -168,22 +168,22 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists views on a bucket.
+ Lists views on a log bucket.
Args:
parent: string, Required. The bucket whose views are to be listed: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" (required)
- pageSize: integer, Optional. The maximum number of results to return from this request. Non-positive values are ignored. The presence of nextPageToken in the response indicates that more results might be available.
+ pageSize: integer, Optional. The maximum number of results to return from this request.Non-positive values are ignored. The presence of nextPageToken in the response indicates that more results might be available.
pageToken: string, Optional. If present, then retrieve the next batch of results from the preceding call to this method. pageToken must be the value of nextPageToken from the previous response. The values of other method parameters should be identical to those in the previous call.
x__xgafv: string, V1 error format.
Allowed values
@@ -196,11 +196,11 @@ Method Details
{ # The response from ListViews.
"nextPageToken": "A String", # If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken.
"views": [ # A list of views.
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
},
],
@@ -223,22 +223,22 @@ Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
- Updates a view. This method replaces the following fields in the existing view with values from the new view: filter.
+ Updates a view on a log bucket. This method replaces the following fields in the existing view with values from the new view: filter.
Args:
- name: string, Required. The full resource name of the view to update "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The full resource name of the view to update "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a view over logs in a bucket.
+{ # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
- updateMask: string, Optional. Field mask that specifies the fields in view that need an update. A field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in view that need an update. A field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -247,11 +247,11 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
diff --git a/docs/dyn/logging_v2.organizations.exclusions.html b/docs/dyn/logging_v2.organizations.exclusions.html
index e55eaa1115a..034c6969c9f 100644
--- a/docs/dyn/logging_v2.organizations.exclusions.html
+++ b/docs/dyn/logging_v2.organizations.exclusions.html
@@ -106,15 +106,15 @@ Method Details
Creates a new exclusion in a specified parent resource. Only log entries belonging to that resource can be excluded. You can have up to 10 exclusions in a resource.
Args:
- parent: string, Required. The parent resource in which to create the exclusion: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" Examples: "projects/my-logging-project", "organizations/123456789". (required)
+ parent: string, Required. The parent resource in which to create the exclusion: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" For examples:"projects/my-logging-project" "organizations/123456789" (required)
body: object, The request body.
The object takes the form of:
-{ # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+{ # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -127,11 +127,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -142,7 +142,7 @@ Method Details
Deletes an exclusion.
Args:
- name: string, Required. The resource name of an existing exclusion to delete: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of an existing exclusion to delete: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -160,7 +160,7 @@ Method Details
Gets the description of an exclusion.
Args:
- name: string, Required. The resource name of an existing exclusion: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of an existing exclusion: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -169,11 +169,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -197,11 +197,11 @@ Method Details
{ # Result returned from ListExclusions.
"exclusions": [ # A list of exclusions.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
@@ -229,15 +229,15 @@ Method Details
Changes one or more properties of an existing exclusion.
Args:
- name: string, Required. The resource name of the exclusion to update: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of the exclusion to update: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
body: object, The request body.
The object takes the form of:
-{ # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+{ # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -251,11 +251,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
diff --git a/docs/dyn/logging_v2.organizations.html b/docs/dyn/logging_v2.organizations.html
index 61f1f9daa53..b9858200260 100644
--- a/docs/dyn/logging_v2.organizations.html
+++ b/docs/dyn/logging_v2.organizations.html
@@ -114,7 +114,7 @@ Method Details
Gets the Logs Router CMEK settings for the given resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
Args:
- name: string, Required. The resource for which to retrieve CMEK settings. "projects/[PROJECT_ID]/cmekSettings" "organizations/[ORGANIZATION_ID]/cmekSettings" "billingAccounts/[BILLING_ACCOUNT_ID]/cmekSettings" "folders/[FOLDER_ID]/cmekSettings" Example: "organizations/12345/cmekSettings".Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization. (required)
+ name: string, Required. The resource for which to retrieve CMEK settings. "projects/[PROJECT_ID]/cmekSettings" "organizations/[ORGANIZATION_ID]/cmekSettings" "billingAccounts/[BILLING_ACCOUNT_ID]/cmekSettings" "folders/[FOLDER_ID]/cmekSettings" For example:"organizations/12345/cmekSettings"Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization. (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -124,7 +124,7 @@ Method Details
An object of the form:
{ # Describes the customer-managed encryption key (CMEK) settings associated with a project, folder, organization, billing account, or flexible resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
- "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/PROJECT_ID/locations/LOCATION/keyRings/KEYRING/cryptoKeys/KEY"For example: "projects/my-project-id/locations/my-region/keyRings/key-ring-name/cryptoKeys/key-name"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
+ "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/[PROJECT_ID]/locations/[LOCATION]/keyRings/[KEYRING]/cryptoKeys/[KEY]" For example:"projects/my-project/locations/us-central1/keyRings/my-ring/cryptoKeys/my-key"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
"name": "A String", # Output only. The resource name of the CMEK settings.
"serviceAccountId": "A String", # Output only. The service account that will be used by the Logs Router to access your Cloud KMS key.Before enabling CMEK for Logs Router, you must first assign the role roles/cloudkms.cryptoKeyEncrypterDecrypter to the service account that the Logs Router will use to access your Cloud KMS key. Use GetCmekSettings to obtain the service account ID.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
}
@@ -135,17 +135,17 @@ Method Details
Updates the Logs Router CMEK settings for the given resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.UpdateCmekSettings will fail if 1) kms_key_name is invalid, or 2) the associated service account does not have the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key, or 3) access to the key is disabled.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
Args:
- name: string, Required. The resource name for the CMEK settings to update. "projects/[PROJECT_ID]/cmekSettings" "organizations/[ORGANIZATION_ID]/cmekSettings" "billingAccounts/[BILLING_ACCOUNT_ID]/cmekSettings" "folders/[FOLDER_ID]/cmekSettings" Example: "organizations/12345/cmekSettings".Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization. (required)
+ name: string, Required. The resource name for the CMEK settings to update. "projects/[PROJECT_ID]/cmekSettings" "organizations/[ORGANIZATION_ID]/cmekSettings" "billingAccounts/[BILLING_ACCOUNT_ID]/cmekSettings" "folders/[FOLDER_ID]/cmekSettings" For example:"organizations/12345/cmekSettings"Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization. (required)
body: object, The request body.
The object takes the form of:
{ # Describes the customer-managed encryption key (CMEK) settings associated with a project, folder, organization, billing account, or flexible resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
- "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/PROJECT_ID/locations/LOCATION/keyRings/KEYRING/cryptoKeys/KEY"For example: "projects/my-project-id/locations/my-region/keyRings/key-ring-name/cryptoKeys/key-name"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
+ "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/[PROJECT_ID]/locations/[LOCATION]/keyRings/[KEYRING]/cryptoKeys/[KEY]" For example:"projects/my-project/locations/us-central1/keyRings/my-ring/cryptoKeys/my-key"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
"name": "A String", # Output only. The resource name of the CMEK settings.
"serviceAccountId": "A String", # Output only. The service account that will be used by the Logs Router to access your Cloud KMS key.Before enabling CMEK for Logs Router, you must first assign the role roles/cloudkms.cryptoKeyEncrypterDecrypter to the service account that the Logs Router will use to access your Cloud KMS key. Use GetCmekSettings to obtain the service account ID.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
}
- updateMask: string, Optional. Field mask identifying which fields from cmek_settings should be updated. A field will be overwritten if and only if it is in the update mask. Output only fields cannot be updated.See FieldMask for more information.Example: "updateMask=kmsKeyName"
+ updateMask: string, Optional. Field mask identifying which fields from cmek_settings should be updated. A field will be overwritten if and only if it is in the update mask. Output only fields cannot be updated.See FieldMask for more information.For example: "updateMask=kmsKeyName"
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -155,7 +155,7 @@ Method Details
An object of the form:
{ # Describes the customer-managed encryption key (CMEK) settings associated with a project, folder, organization, billing account, or flexible resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
- "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/PROJECT_ID/locations/LOCATION/keyRings/KEYRING/cryptoKeys/KEY"For example: "projects/my-project-id/locations/my-region/keyRings/key-ring-name/cryptoKeys/key-name"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
+ "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/[PROJECT_ID]/locations/[LOCATION]/keyRings/[KEYRING]/cryptoKeys/[KEY]" For example:"projects/my-project/locations/us-central1/keyRings/my-ring/cryptoKeys/my-key"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
"name": "A String", # Output only. The resource name of the CMEK settings.
"serviceAccountId": "A String", # Output only. The service account that will be used by the Logs Router to access your Cloud KMS key.Before enabling CMEK for Logs Router, you must first assign the role roles/cloudkms.cryptoKeyEncrypterDecrypter to the service account that the Logs Router will use to access your Cloud KMS key. Use GetCmekSettings to obtain the service account ID.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
}
diff --git a/docs/dyn/logging_v2.organizations.locations.buckets.html b/docs/dyn/logging_v2.organizations.locations.buckets.html
index 5ec17bb5ed4..327af515b76 100644
--- a/docs/dyn/logging_v2.organizations.locations.buckets.html
+++ b/docs/dyn/logging_v2.organizations.locations.buckets.html
@@ -84,25 +84,25 @@ Instance Methods
Close httplib2 connections.
create(parent, body=None, bucketId=None, x__xgafv=None)
-Creates a bucket that can be used to store log entries. Once a bucket has been created, the region cannot be changed.
+Creates a log bucket that can be used to store log entries. After a bucket has been created, the bucket's location cannot be changed.
-Deletes a bucket. Moves the bucket to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all logs in the bucket will be permanently deleted.
+Deletes a log bucket.Changes the bucket's lifecycle_state to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all log entries in the bucket will be permanently deleted.
-Gets a bucket.
+Gets a log bucket.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists buckets.
+Lists log buckets.
list_next(previous_request, previous_response)
Retrieves the next page of results.
patch(name, body=None, updateMask=None, x__xgafv=None)
-Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created.
+Updates a log bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a lifecycle_state of DELETE_REQUESTED, then FAILED_PRECONDITION will be returned.After a bucket has been created, the bucket's location cannot be changed.
undelete(name, body=None, x__xgafv=None)
-Undeletes a bucket. A bucket that has been deleted may be undeleted within the grace period of 7 days.
+Undeletes a log bucket. A bucket that has been deleted can be undeleted within the grace period of 7 days.
Method Details
close()
@@ -111,20 +111,20 @@ Method Details
create(parent, body=None, bucketId=None, x__xgafv=None)
- Creates a bucket that can be used to store log entries. Once a bucket has been created, the region cannot be changed.
+ Creates a log bucket that can be used to store log entries. After a bucket has been created, the bucket's location cannot be changed.
Args:
- parent: string, Required. The resource in which to create the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" Example: "projects/my-logging-project/locations/global" (required)
+ parent: string, Required. The resource in which to create the log bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" For example:"projects/my-project/locations/global" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a repository of logs.
+{ # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -140,13 +140,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -156,10 +156,10 @@ Method Details
delete(name, x__xgafv=None)
- Deletes a bucket. Moves the bucket to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all logs in the bucket will be permanently deleted.
+ Deletes a log bucket.Changes the bucket's lifecycle_state to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all log entries in the bucket will be permanently deleted.
Args:
- name: string, Required. The full resource name of the bucket to delete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The full resource name of the bucket to delete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -174,10 +174,10 @@ Method Details
get(name, x__xgafv=None)
- Gets a bucket.
+ Gets a log bucket.
Args:
- name: string, Required. The resource name of the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The resource name of the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -186,13 +186,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -202,7 +202,7 @@ Method Details
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists buckets.
+ Lists log buckets.
Args:
parent: string, Required. The parent resource whose buckets are to be listed: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]" Note: The locations portion of the resource must be specified, but supplying the character - in place of LOCATION_ID will return all buckets. (required)
@@ -218,13 +218,13 @@ Method Details
{ # The response from ListBuckets.
"buckets": [ # A list of buckets.
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -251,27 +251,27 @@ Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
- Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created.
+ Updates a log bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a lifecycle_state of DELETE_REQUESTED, then FAILED_PRECONDITION will be returned.After a bucket has been created, the bucket's location cannot be changed.
Args:
- name: string, Required. The full resource name of the bucket to update. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". Also requires permission "resourcemanager.projects.updateLiens" to set the locked property (required)
+ name: string, Required. The full resource name of the bucket to update. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a repository of logs.
+{ # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
"updateTime": "A String", # Output only. The last update timestamp of the bucket.
}
- updateMask: string, Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=retention_days.
+ updateMask: string, Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see: https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=retention_days
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -280,13 +280,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -296,10 +296,10 @@ Method Details
undelete(name, body=None, x__xgafv=None)
- Undeletes a bucket. A bucket that has been deleted may be undeleted within the grace period of 7 days.
+ Undeletes a log bucket. A bucket that has been deleted can be undeleted within the grace period of 7 days.
Args:
- name: string, Required. The full resource name of the bucket to undelete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The full resource name of the bucket to undelete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
diff --git a/docs/dyn/logging_v2.organizations.locations.buckets.views.html b/docs/dyn/logging_v2.organizations.locations.buckets.views.html
index bebdc040c5b..770d83bc458 100644
--- a/docs/dyn/logging_v2.organizations.locations.buckets.views.html
+++ b/docs/dyn/logging_v2.organizations.locations.buckets.views.html
@@ -79,22 +79,22 @@ Instance Methods
Close httplib2 connections.
create(parent, body=None, viewId=None, x__xgafv=None)
-Creates a view over logs in a bucket. A bucket may contain a maximum of 50 views.
+Creates a view over log entries in a log bucket. A bucket may contain a maximum of 50 views.
-Deletes a view from a bucket.
+Deletes a view on a log bucket.
-Gets a view.
+Gets a view on a log bucket..
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists views on a bucket.
+Lists views on a log bucket.
list_next(previous_request, previous_response)
Retrieves the next page of results.
patch(name, body=None, updateMask=None, x__xgafv=None)
-Updates a view. This method replaces the following fields in the existing view with values from the new view: filter.
+Updates a view on a log bucket. This method replaces the following fields in the existing view with values from the new view: filter.
Method Details
close()
@@ -103,18 +103,18 @@ Method Details
create(parent, body=None, viewId=None, x__xgafv=None)
- Creates a view over logs in a bucket. A bucket may contain a maximum of 50 views.
+ Creates a view over log entries in a log bucket. A bucket may contain a maximum of 50 views.
Args:
- parent: string, Required. The bucket in which to create the view "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-logging-project/locations/my-location/buckets/my-bucket" (required)
+ parent: string, Required. The bucket in which to create the view `"projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]"` For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a view over logs in a bucket.
+{ # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
@@ -127,21 +127,21 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
delete(name, x__xgafv=None)
- Deletes a view from a bucket.
+ Deletes a view on a log bucket.
Args:
- name: string, Required. The full resource name of the view to delete: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The full resource name of the view to delete: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -156,10 +156,10 @@ Method Details
get(name, x__xgafv=None)
- Gets a view.
+ Gets a view on a log bucket..
Args:
- name: string, Required. The resource name of the policy: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The resource name of the policy: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -168,22 +168,22 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists views on a bucket.
+ Lists views on a log bucket.
Args:
parent: string, Required. The bucket whose views are to be listed: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" (required)
- pageSize: integer, Optional. The maximum number of results to return from this request. Non-positive values are ignored. The presence of nextPageToken in the response indicates that more results might be available.
+ pageSize: integer, Optional. The maximum number of results to return from this request.Non-positive values are ignored. The presence of nextPageToken in the response indicates that more results might be available.
pageToken: string, Optional. If present, then retrieve the next batch of results from the preceding call to this method. pageToken must be the value of nextPageToken from the previous response. The values of other method parameters should be identical to those in the previous call.
x__xgafv: string, V1 error format.
Allowed values
@@ -196,11 +196,11 @@ Method Details
{ # The response from ListViews.
"nextPageToken": "A String", # If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken.
"views": [ # A list of views.
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
},
],
@@ -223,22 +223,22 @@ Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
- Updates a view. This method replaces the following fields in the existing view with values from the new view: filter.
+ Updates a view on a log bucket. This method replaces the following fields in the existing view with values from the new view: filter.
Args:
- name: string, Required. The full resource name of the view to update "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The full resource name of the view to update "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a view over logs in a bucket.
+{ # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
- updateMask: string, Optional. Field mask that specifies the fields in view that need an update. A field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in view that need an update. A field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -247,11 +247,11 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
diff --git a/docs/dyn/logging_v2.organizations.sinks.html b/docs/dyn/logging_v2.organizations.sinks.html
index d2327429d5b..fec66f8088f 100644
--- a/docs/dyn/logging_v2.organizations.sinks.html
+++ b/docs/dyn/logging_v2.organizations.sinks.html
@@ -109,38 +109,38 @@ Method Details
Creates a sink that exports specified log entries to a destination. The export of newly-ingested log entries begins immediately, unless the sink's writer_identity is not permitted to write to the destination. A sink can export log entries only from the resource owning the sink.
Args:
- parent: string, Required. The resource in which to create the sink: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" Examples: "projects/my-logging-project", "organizations/123456789". (required)
+ parent: string, Required. The resource in which to create the sink: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" For examples:"projects/my-project" "organizations/123456789" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
- uniqueWriterIdentity: boolean, Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.
+ uniqueWriterIdentity: boolean, Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Cloud Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -149,31 +149,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -182,7 +182,7 @@ Method Details
Deletes a sink. If the sink has a unique writer_identity, then that service account is also deleted.
Args:
- sinkName: string, Required. The full resource name of the sink to delete, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to delete, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -200,7 +200,7 @@ Method Details
Gets a sink.
Args:
- sinkName: string, Required. The resource name of the sink: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The resource name of the sink: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -209,31 +209,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -256,31 +256,31 @@ Method Details
{ # Result returned from ListSinks.
"nextPageToken": "A String", # If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken.
"sinks": [ # A list of sinks.
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
},
],
}
@@ -305,39 +305,39 @@ Method Details
Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.
Args:
- sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
uniqueWriterIdentity: boolean, Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field: If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity. If the old value is false and the new value is true, then writer_identity is changed to a unique service account. It is an error if the old value is true and the new value is set to false or defaulted to false.
- updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes:destination,filter,includeChildrenAt some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -346,31 +346,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -379,39 +379,39 @@ Method Details
Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.
Args:
- sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
uniqueWriterIdentity: boolean, Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field: If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity. If the old value is false and the new value is true, then writer_identity is changed to a unique service account. It is an error if the old value is true and the new value is set to false or defaulted to false.
- updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes:destination,filter,includeChildrenAt some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -420,31 +420,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
diff --git a/docs/dyn/logging_v2.projects.exclusions.html b/docs/dyn/logging_v2.projects.exclusions.html
index 36053e2283a..758b5394d45 100644
--- a/docs/dyn/logging_v2.projects.exclusions.html
+++ b/docs/dyn/logging_v2.projects.exclusions.html
@@ -106,15 +106,15 @@ Method Details
Creates a new exclusion in a specified parent resource. Only log entries belonging to that resource can be excluded. You can have up to 10 exclusions in a resource.
Args:
- parent: string, Required. The parent resource in which to create the exclusion: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" Examples: "projects/my-logging-project", "organizations/123456789". (required)
+ parent: string, Required. The parent resource in which to create the exclusion: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" For examples:"projects/my-logging-project" "organizations/123456789" (required)
body: object, The request body.
The object takes the form of:
-{ # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+{ # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -127,11 +127,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -142,7 +142,7 @@ Method Details
Deletes an exclusion.
Args:
- name: string, Required. The resource name of an existing exclusion to delete: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of an existing exclusion to delete: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -160,7 +160,7 @@ Method Details
Gets the description of an exclusion.
Args:
- name: string, Required. The resource name of an existing exclusion: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of an existing exclusion: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -169,11 +169,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -197,11 +197,11 @@ Method Details
{ # Result returned from ListExclusions.
"exclusions": [ # A list of exclusions.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
@@ -229,15 +229,15 @@ Method Details
Changes one or more properties of an existing exclusion.
Args:
- name: string, Required. The resource name of the exclusion to update: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" Example: "projects/my-project-id/exclusions/my-exclusion-id". (required)
+ name: string, Required. The resource name of the exclusion to update: "projects/[PROJECT_ID]/exclusions/[EXCLUSION_ID]" "organizations/[ORGANIZATION_ID]/exclusions/[EXCLUSION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/exclusions/[EXCLUSION_ID]" "folders/[FOLDER_ID]/exclusions/[EXCLUSION_ID]" For example:"projects/my-project/exclusions/my-exclusion" (required)
body: object, The request body.
The object takes the form of:
-{ # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+{ # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
@@ -251,11 +251,11 @@ Method Details
Returns:
An object of the form:
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
}
diff --git a/docs/dyn/logging_v2.projects.locations.buckets.html b/docs/dyn/logging_v2.projects.locations.buckets.html
index 1bc9c1054c1..74186b35e75 100644
--- a/docs/dyn/logging_v2.projects.locations.buckets.html
+++ b/docs/dyn/logging_v2.projects.locations.buckets.html
@@ -84,25 +84,25 @@ Instance Methods
Close httplib2 connections.
create(parent, body=None, bucketId=None, x__xgafv=None)
-Creates a bucket that can be used to store log entries. Once a bucket has been created, the region cannot be changed.
+Creates a log bucket that can be used to store log entries. After a bucket has been created, the bucket's location cannot be changed.
-Deletes a bucket. Moves the bucket to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all logs in the bucket will be permanently deleted.
+Deletes a log bucket.Changes the bucket's lifecycle_state to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all log entries in the bucket will be permanently deleted.
-Gets a bucket.
+Gets a log bucket.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists buckets.
+Lists log buckets.
list_next(previous_request, previous_response)
Retrieves the next page of results.
patch(name, body=None, updateMask=None, x__xgafv=None)
-Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created.
+Updates a log bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a lifecycle_state of DELETE_REQUESTED, then FAILED_PRECONDITION will be returned.After a bucket has been created, the bucket's location cannot be changed.
undelete(name, body=None, x__xgafv=None)
-Undeletes a bucket. A bucket that has been deleted may be undeleted within the grace period of 7 days.
+Undeletes a log bucket. A bucket that has been deleted can be undeleted within the grace period of 7 days.
Method Details
close()
@@ -111,20 +111,20 @@ Method Details
create(parent, body=None, bucketId=None, x__xgafv=None)
- Creates a bucket that can be used to store log entries. Once a bucket has been created, the region cannot be changed.
+ Creates a log bucket that can be used to store log entries. After a bucket has been created, the bucket's location cannot be changed.
Args:
- parent: string, Required. The resource in which to create the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" Example: "projects/my-logging-project/locations/global" (required)
+ parent: string, Required. The resource in which to create the log bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" For example:"projects/my-project/locations/global" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a repository of logs.
+{ # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -140,13 +140,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -156,10 +156,10 @@ Method Details
delete(name, x__xgafv=None)
- Deletes a bucket. Moves the bucket to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all logs in the bucket will be permanently deleted.
+ Deletes a log bucket.Changes the bucket's lifecycle_state to the DELETE_REQUESTED state. After 7 days, the bucket will be purged and all log entries in the bucket will be permanently deleted.
Args:
- name: string, Required. The full resource name of the bucket to delete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The full resource name of the bucket to delete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -174,10 +174,10 @@ Method Details
get(name, x__xgafv=None)
- Gets a bucket.
+ Gets a log bucket.
Args:
- name: string, Required. The resource name of the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The resource name of the bucket: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -186,13 +186,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -202,7 +202,7 @@ Method Details
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists buckets.
+ Lists log buckets.
Args:
parent: string, Required. The parent resource whose buckets are to be listed: "projects/[PROJECT_ID]/locations/[LOCATION_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]" Note: The locations portion of the resource must be specified, but supplying the character - in place of LOCATION_ID will return all buckets. (required)
@@ -218,13 +218,13 @@ Method Details
{ # The response from ListBuckets.
"buckets": [ # A list of buckets.
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -251,27 +251,27 @@ Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
- Updates a bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a LifecycleState of DELETE_REQUESTED, FAILED_PRECONDITION will be returned.A buckets region may not be modified after it is created.
+ Updates a log bucket. This method replaces the following fields in the existing bucket with values from the new bucket: retention_periodIf the retention period is decreased and the bucket is locked, FAILED_PRECONDITION will be returned.If the bucket has a lifecycle_state of DELETE_REQUESTED, then FAILED_PRECONDITION will be returned.After a bucket has been created, the bucket's location cannot be changed.
Args:
- name: string, Required. The full resource name of the bucket to update. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". Also requires permission "resourcemanager.projects.updateLiens" to set the locked property (required)
+ name: string, Required. The full resource name of the bucket to update. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a repository of logs.
+{ # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
"updateTime": "A String", # Output only. The last update timestamp of the bucket.
}
- updateMask: string, Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=retention_days.
+ updateMask: string, Required. Field mask that specifies the fields in bucket that need an update. A bucket field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see: https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=retention_days
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -280,13 +280,13 @@ Method Details
Returns:
An object of the form:
- { # Describes a repository of logs.
+ { # Describes a repository in which log entries are stored.
"createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets.
"description": "A String", # Describes this bucket.
"lifecycleState": "A String", # Output only. The bucket lifecycle state.
- "locked": True or False, # Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty.
- "name": "A String", # Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed.
- "restrictedFields": [ # Log entry field paths that are denied access in this bucket. The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation. Restricting a repeated field will restrict all values. Adding a parent will block all child fields e.g. foo.bar will block foo.bar.baz.
+ "locked": True or False, # Whether the bucket is locked.The retention period on a locked bucket cannot be changed. Locked buckets may only be deleted if they are empty.
+ "name": "A String", # Output only. The resource name of the bucket.For example:projects/my-project/locations/global/buckets/my-bucketFor a list of supported locations, see Supported Regions (https://cloud.google.com/logging/docs/region-support)For the location of global it is unspecified where log entries are actually stored.After a bucket has been created, the location cannot be changed.
+ "restrictedFields": [ # Log entry field paths that are denied access in this bucket.The following fields and their children are eligible: textPayload, jsonPayload, protoPayload, httpRequest, labels, sourceLocation.Restricting a repeated field will restrict all values. Adding a parent will block all child fields. (e.g. foo.bar will block foo.bar.baz)
"A String",
],
"retentionDays": 42, # Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used.
@@ -296,10 +296,10 @@ Method Details
undelete(name, body=None, x__xgafv=None)
- Undeletes a bucket. A bucket that has been deleted may be undeleted within the grace period of 7 days.
+ Undeletes a log bucket. A bucket that has been deleted can be undeleted within the grace period of 7 days.
Args:
- name: string, Required. The full resource name of the bucket to undelete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id". (required)
+ name: string, Required. The full resource name of the bucket to undelete. "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "organizations/[ORGANIZATION_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" "folders/[FOLDER_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
diff --git a/docs/dyn/logging_v2.projects.locations.buckets.views.html b/docs/dyn/logging_v2.projects.locations.buckets.views.html
index 8131411f240..0f826e5feb8 100644
--- a/docs/dyn/logging_v2.projects.locations.buckets.views.html
+++ b/docs/dyn/logging_v2.projects.locations.buckets.views.html
@@ -79,22 +79,22 @@ Instance Methods
Close httplib2 connections.
create(parent, body=None, viewId=None, x__xgafv=None)
-Creates a view over logs in a bucket. A bucket may contain a maximum of 50 views.
+Creates a view over log entries in a log bucket. A bucket may contain a maximum of 50 views.
-Deletes a view from a bucket.
+Deletes a view on a log bucket.
-Gets a view.
+Gets a view on a log bucket..
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists views on a bucket.
+Lists views on a log bucket.
list_next(previous_request, previous_response)
Retrieves the next page of results.
patch(name, body=None, updateMask=None, x__xgafv=None)
-Updates a view. This method replaces the following fields in the existing view with values from the new view: filter.
+Updates a view on a log bucket. This method replaces the following fields in the existing view with values from the new view: filter.
Method Details
close()
@@ -103,18 +103,18 @@ Method Details
create(parent, body=None, viewId=None, x__xgafv=None)
- Creates a view over logs in a bucket. A bucket may contain a maximum of 50 views.
+ Creates a view over log entries in a log bucket. A bucket may contain a maximum of 50 views.
Args:
- parent: string, Required. The bucket in which to create the view "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" Example: "projects/my-logging-project/locations/my-location/buckets/my-bucket" (required)
+ parent: string, Required. The bucket in which to create the view `"projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]"` For example:"projects/my-project/locations/global/buckets/my-bucket" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a view over logs in a bucket.
+{ # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
@@ -127,21 +127,21 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
delete(name, x__xgafv=None)
- Deletes a view from a bucket.
+ Deletes a view on a log bucket.
Args:
- name: string, Required. The full resource name of the view to delete: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The full resource name of the view to delete: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -156,10 +156,10 @@ Method Details
get(name, x__xgafv=None)
- Gets a view.
+ Gets a view on a log bucket..
Args:
- name: string, Required. The resource name of the policy: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The resource name of the policy: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -168,22 +168,22 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists views on a bucket.
+ Lists views on a log bucket.
Args:
parent: string, Required. The bucket whose views are to be listed: "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]" (required)
- pageSize: integer, Optional. The maximum number of results to return from this request. Non-positive values are ignored. The presence of nextPageToken in the response indicates that more results might be available.
+ pageSize: integer, Optional. The maximum number of results to return from this request.Non-positive values are ignored. The presence of nextPageToken in the response indicates that more results might be available.
pageToken: string, Optional. If present, then retrieve the next batch of results from the preceding call to this method. pageToken must be the value of nextPageToken from the previous response. The values of other method parameters should be identical to those in the previous call.
x__xgafv: string, V1 error format.
Allowed values
@@ -196,11 +196,11 @@ Method Details
{ # The response from ListViews.
"nextPageToken": "A String", # If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken.
"views": [ # A list of views.
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
},
],
@@ -223,22 +223,22 @@ Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
- Updates a view. This method replaces the following fields in the existing view with values from the new view: filter.
+ Updates a view on a log bucket. This method replaces the following fields in the existing view with values from the new view: filter.
Args:
- name: string, Required. The full resource name of the view to update "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" Example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view-id". (required)
+ name: string, Required. The full resource name of the view to update "projects/[PROJECT_ID]/locations/[LOCATION_ID]/buckets/[BUCKET_ID]/views/[VIEW_ID]" For example:"projects/my-project/locations/global/buckets/my-bucket/views/my-view" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a view over logs in a bucket.
+{ # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
- updateMask: string, Optional. Field mask that specifies the fields in view that need an update. A field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in view that need an update. A field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -247,11 +247,11 @@ Method Details
Returns:
An object of the form:
- { # Describes a view over logs in a bucket.
+ { # Describes a view over log entries in a bucket.
"createTime": "A String", # Output only. The creation timestamp of the view.
"description": "A String", # Describes this view.
- "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
- "name": "A String", # The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view
+ "filter": "A String", # Filter that restricts which log entries in a bucket are visible in this view.Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log idFor example:SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout")
+ "name": "A String", # The resource name of the view.For example:projects/my-project/locations/global/buckets/my-bucket/views/my-view
"updateTime": "A String", # Output only. The last update timestamp of the view.
}
diff --git a/docs/dyn/logging_v2.projects.sinks.html b/docs/dyn/logging_v2.projects.sinks.html
index e26e21255aa..6f56c4b414a 100644
--- a/docs/dyn/logging_v2.projects.sinks.html
+++ b/docs/dyn/logging_v2.projects.sinks.html
@@ -109,38 +109,38 @@ Method Details
Creates a sink that exports specified log entries to a destination. The export of newly-ingested log entries begins immediately, unless the sink's writer_identity is not permitted to write to the destination. A sink can export log entries only from the resource owning the sink.
Args:
- parent: string, Required. The resource in which to create the sink: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" Examples: "projects/my-logging-project", "organizations/123456789". (required)
+ parent: string, Required. The resource in which to create the sink: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" For examples:"projects/my-project" "organizations/123456789" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
- uniqueWriterIdentity: boolean, Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.
+ uniqueWriterIdentity: boolean, Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Cloud Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -149,31 +149,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -182,7 +182,7 @@ Method Details
Deletes a sink. If the sink has a unique writer_identity, then that service account is also deleted.
Args:
- sinkName: string, Required. The full resource name of the sink to delete, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to delete, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -200,7 +200,7 @@ Method Details
Gets a sink.
Args:
- sinkName: string, Required. The resource name of the sink: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The resource name of the sink: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -209,31 +209,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -256,31 +256,31 @@ Method Details
{ # Result returned from ListSinks.
"nextPageToken": "A String", # If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken.
"sinks": [ # A list of sinks.
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
},
],
}
@@ -305,39 +305,39 @@ Method Details
Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.
Args:
- sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
uniqueWriterIdentity: boolean, Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field: If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity. If the old value is false and the new value is true, then writer_identity is changed to a unique service account. It is an error if the old value is true and the new value is set to false or defaulted to false.
- updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes:destination,filter,includeChildrenAt some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -346,31 +346,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -379,39 +379,39 @@ Method Details
Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.
Args:
- sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
uniqueWriterIdentity: boolean, Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field: If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity. If the old value is false and the new value is true, then writer_identity is changed to a unique service account. It is an error if the old value is true and the new value is set to false or defaulted to false.
- updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes:destination,filter,includeChildrenAt some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -420,31 +420,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
diff --git a/docs/dyn/logging_v2.sinks.html b/docs/dyn/logging_v2.sinks.html
index cc15f37c8e8..9f68f54d983 100644
--- a/docs/dyn/logging_v2.sinks.html
+++ b/docs/dyn/logging_v2.sinks.html
@@ -106,38 +106,38 @@ Method Details
Creates a sink that exports specified log entries to a destination. The export of newly-ingested log entries begins immediately, unless the sink's writer_identity is not permitted to write to the destination. A sink can export log entries only from the resource owning the sink.
Args:
- parent: string, Required. The resource in which to create the sink: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" Examples: "projects/my-logging-project", "organizations/123456789". (required)
+ parent: string, Required. The resource in which to create the sink: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" For examples:"projects/my-project" "organizations/123456789" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
- uniqueWriterIdentity: boolean, Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.
+ uniqueWriterIdentity: boolean, Optional. Determines the kind of IAM identity returned as writer_identity in the new sink. If this value is omitted or set to false, and if the sink's parent is a project, then the value returned as writer_identity is the same group or service account used by Cloud Logging before the addition of writer identities to this API. The sink's destination must be in the same project as the sink itself.If this field is set to true, or if the sink is owned by a non-project resource such as an organization, then the value of writer_identity will be a unique service account used only for exports from the new sink. For more information, see writer_identity in LogSink.
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -146,31 +146,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -179,7 +179,7 @@ Method Details
Deletes a sink. If the sink has a unique writer_identity, then that service account is also deleted.
Args:
- sinkName: string, Required. The full resource name of the sink to delete, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to delete, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -197,7 +197,7 @@ Method Details
Gets a sink.
Args:
- sinkName: string, Required. The resource name of the sink: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The resource name of the sink: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -206,31 +206,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
@@ -253,31 +253,31 @@ Method Details
{ # Result returned from ListSinks.
"nextPageToken": "A String", # If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken.
"sinks": [ # A list of sinks.
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
},
],
}
@@ -302,39 +302,39 @@ Method Details
Updates a sink. This method replaces the following fields in the existing sink with values from the new sink: destination, and filter.The updated sink might also have a new writer_identity; see the unique_writer_identity field.
Args:
- sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" Example: "projects/my-project-id/sinks/my-sink-id". (required)
+ sinkName: string, Required. The full resource name of the sink to update, including the parent resource and the sink identifier: "projects/[PROJECT_ID]/sinks/[SINK_ID]" "organizations/[ORGANIZATION_ID]/sinks/[SINK_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/sinks/[SINK_ID]" "folders/[FOLDER_ID]/sinks/[SINK_ID]" For example:"projects/my-project/sinks/my-sink" (required)
body: object, The request body.
The object takes the form of:
-{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+{ # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
uniqueWriterIdentity: boolean, Optional. See sinks.create for a description of this field. When updating a sink, the effect of this field on the value of writer_identity in the updated sink depends on both the old and new values of this field: If the old and new values of this field are both false or both true, then there is no change to the sink's writer_identity. If the old value is false and the new value is true, then writer_identity is changed to a unique service account. It is an error if the old value is true and the new value is set to false or defaulted to false.
- updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes: destination,filter,includeChildren At some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskExample: updateMask=filter.
+ updateMask: string, Optional. Field mask that specifies the fields in sink that need an update. A sink field will be overwritten if, and only if, it is in the update mask. name and output only fields cannot be updated.An empty updateMask is temporarily treated as using the following mask for backwards compatibility purposes:destination,filter,includeChildrenAt some point in the future, behavior will be removed and specifying an empty updateMask will be an error.For a detailed FieldMask definition, see https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#google.protobuf.FieldMaskFor example: updateMask=filter
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -343,31 +343,31 @@ Method Details
Returns:
An object of the form:
- { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
+ { # Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Pub/Sub topic or a Cloud Logging log bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder.
"bigqueryOptions": { # Options that change functionality of a sink exporting data to BigQuery. # Optional. Options that affect sinks exporting data to BigQuery.
- "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
- "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
+ "usePartitionedTables": True or False, # Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Cloud Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone.
+ "usesTimestampColumnPartitioning": True or False, # Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use.All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false.
},
"createTime": "A String", # Output only. The creation timestamp of the sink.This field may not be present for older sinks.
- "description": "A String", # Optional. A description of this sink. The maximum length of the description is 8000 characters.
+ "description": "A String", # Optional. A description of this sink.The maximum length of the description is 8000 characters.
"destination": "A String", # Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs).
- "disabled": True or False, # Optional. If set to True, then this sink is disabled and it does not export any log entries.
- "exclusions": [ # Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
- { # Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
+ "disabled": True or False, # Optional. If set to true, then this sink is disabled and it does not export any log entries.
+ "exclusions": [ # Optional. Log entries that match any of these exclusion filters will not be exported.If a log entry is matched by both filter and one of exclusion_filters it will not be exported.
+ { # Specifies a set of log entries that are not to be stored in Cloud Logging. If your GCP resource receives a large volume of log entries, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries.
"createTime": "A String", # Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions.
"description": "A String", # Optional. A description of this exclusion.
"disabled": True or False, # Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field.
- "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)"
+ "filter": "A String", # Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries.For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)
"name": "A String", # Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric.
"updateTime": "A String", # Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions.
},
],
- "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
- "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
- "name": "A String", # Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
+ "filter": "A String", # Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter.For example:logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR
+ "includeChildren": True or False, # Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then log entries from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression.For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent.To only export entries from certain child projects, filter on the project part of the log name:logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance
+ "name": "A String", # Required. The client-assigned sink identifier, unique within the project.For example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric.
"outputVersionFormat": "A String", # Deprecated. This field is unused.
"updateTime": "A String", # Output only. The last update timestamp of the sink.This field may not be present for older sinks.
- "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
+ "writerIdentity": "A String", # Output only. An IAM identity—a service account or group—under which Cloud Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity.
}
diff --git a/docs/dyn/logging_v2.v2.html b/docs/dyn/logging_v2.v2.html
index d6c6b12d38c..ac125c6e3de 100644
--- a/docs/dyn/logging_v2.v2.html
+++ b/docs/dyn/logging_v2.v2.html
@@ -94,7 +94,7 @@ Method Details
Gets the Logs Router CMEK settings for the given resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
Args:
- name: string, Required. The resource for which to retrieve CMEK settings. "projects/[PROJECT_ID]/cmekSettings" "organizations/[ORGANIZATION_ID]/cmekSettings" "billingAccounts/[BILLING_ACCOUNT_ID]/cmekSettings" "folders/[FOLDER_ID]/cmekSettings" Example: "organizations/12345/cmekSettings".Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization. (required)
+ name: string, Required. The resource for which to retrieve CMEK settings. "projects/[PROJECT_ID]/cmekSettings" "organizations/[ORGANIZATION_ID]/cmekSettings" "billingAccounts/[BILLING_ACCOUNT_ID]/cmekSettings" "folders/[FOLDER_ID]/cmekSettings" For example:"organizations/12345/cmekSettings"Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization. (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -104,7 +104,7 @@ Method Details
An object of the form:
{ # Describes the customer-managed encryption key (CMEK) settings associated with a project, folder, organization, billing account, or flexible resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
- "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/PROJECT_ID/locations/LOCATION/keyRings/KEYRING/cryptoKeys/KEY"For example: "projects/my-project-id/locations/my-region/keyRings/key-ring-name/cryptoKeys/key-name"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
+ "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/[PROJECT_ID]/locations/[LOCATION]/keyRings/[KEYRING]/cryptoKeys/[KEY]" For example:"projects/my-project/locations/us-central1/keyRings/my-ring/cryptoKeys/my-key"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
"name": "A String", # Output only. The resource name of the CMEK settings.
"serviceAccountId": "A String", # Output only. The service account that will be used by the Logs Router to access your Cloud KMS key.Before enabling CMEK for Logs Router, you must first assign the role roles/cloudkms.cryptoKeyEncrypterDecrypter to the service account that the Logs Router will use to access your Cloud KMS key. Use GetCmekSettings to obtain the service account ID.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
}
@@ -115,17 +115,17 @@ Method Details
Updates the Logs Router CMEK settings for the given resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.UpdateCmekSettings will fail if 1) kms_key_name is invalid, or 2) the associated service account does not have the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key, or 3) access to the key is disabled.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
Args:
- name: string, Required. The resource name for the CMEK settings to update. "projects/[PROJECT_ID]/cmekSettings" "organizations/[ORGANIZATION_ID]/cmekSettings" "billingAccounts/[BILLING_ACCOUNT_ID]/cmekSettings" "folders/[FOLDER_ID]/cmekSettings" Example: "organizations/12345/cmekSettings".Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization. (required)
+ name: string, Required. The resource name for the CMEK settings to update. "projects/[PROJECT_ID]/cmekSettings" "organizations/[ORGANIZATION_ID]/cmekSettings" "billingAccounts/[BILLING_ACCOUNT_ID]/cmekSettings" "folders/[FOLDER_ID]/cmekSettings" For example:"organizations/12345/cmekSettings"Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization. (required)
body: object, The request body.
The object takes the form of:
{ # Describes the customer-managed encryption key (CMEK) settings associated with a project, folder, organization, billing account, or flexible resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
- "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/PROJECT_ID/locations/LOCATION/keyRings/KEYRING/cryptoKeys/KEY"For example: "projects/my-project-id/locations/my-region/keyRings/key-ring-name/cryptoKeys/key-name"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
+ "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/[PROJECT_ID]/locations/[LOCATION]/keyRings/[KEYRING]/cryptoKeys/[KEY]" For example:"projects/my-project/locations/us-central1/keyRings/my-ring/cryptoKeys/my-key"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
"name": "A String", # Output only. The resource name of the CMEK settings.
"serviceAccountId": "A String", # Output only. The service account that will be used by the Logs Router to access your Cloud KMS key.Before enabling CMEK for Logs Router, you must first assign the role roles/cloudkms.cryptoKeyEncrypterDecrypter to the service account that the Logs Router will use to access your Cloud KMS key. Use GetCmekSettings to obtain the service account ID.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
}
- updateMask: string, Optional. Field mask identifying which fields from cmek_settings should be updated. A field will be overwritten if and only if it is in the update mask. Output only fields cannot be updated.See FieldMask for more information.Example: "updateMask=kmsKeyName"
+ updateMask: string, Optional. Field mask identifying which fields from cmek_settings should be updated. A field will be overwritten if and only if it is in the update mask. Output only fields cannot be updated.See FieldMask for more information.For example: "updateMask=kmsKeyName"
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -135,7 +135,7 @@ Method Details
An object of the form:
{ # Describes the customer-managed encryption key (CMEK) settings associated with a project, folder, organization, billing account, or flexible resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
- "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/PROJECT_ID/locations/LOCATION/keyRings/KEYRING/cryptoKeys/KEY"For example: "projects/my-project-id/locations/my-region/keyRings/key-ring-name/cryptoKeys/key-name"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
+ "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/[PROJECT_ID]/locations/[LOCATION]/keyRings/[KEYRING]/cryptoKeys/[KEY]" For example:"projects/my-project/locations/us-central1/keyRings/my-ring/cryptoKeys/my-key"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
"name": "A String", # Output only. The resource name of the CMEK settings.
"serviceAccountId": "A String", # Output only. The service account that will be used by the Logs Router to access your Cloud KMS key.Before enabling CMEK for Logs Router, you must first assign the role roles/cloudkms.cryptoKeyEncrypterDecrypter to the service account that the Logs Router will use to access your Cloud KMS key. Use GetCmekSettings to obtain the service account ID.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
}
diff --git a/docs/dyn/metastore_v1alpha.projects.locations.services.backups.html b/docs/dyn/metastore_v1alpha.projects.locations.services.backups.html
index e870ec443ba..f5410422935 100644
--- a/docs/dyn/metastore_v1alpha.projects.locations.services.backups.html
+++ b/docs/dyn/metastore_v1alpha.projects.locations.services.backups.html
@@ -121,6 +121,9 @@ Method Details
"description": "A String", # The description of the backup.
"endTime": "A String", # Output only. The time when the backup finished creating.
"name": "A String", # Immutable. The relative resource name of the backup, in the following form:projects/{project_number}/locations/{location_id}/services/{service_id}/backups/{backup_id}
+ "restoringServices": [ # Output only. Services that are restoring from the backup.
+ "A String",
+ ],
"serviceRevision": { # A managed metastore service that serves metadata queries. # Output only. The revision of the service at the time of backup.
"artifactGcsUri": "A String", # Output only. A Cloud Storage URI (starting with gs://) that specifies where artifacts related to the metastore service are stored.
"createTime": "A String", # Output only. The time when the metastore service was created.
@@ -273,6 +276,9 @@ Method Details
"description": "A String", # The description of the backup.
"endTime": "A String", # Output only. The time when the backup finished creating.
"name": "A String", # Immutable. The relative resource name of the backup, in the following form:projects/{project_number}/locations/{location_id}/services/{service_id}/backups/{backup_id}
+ "restoringServices": [ # Output only. Services that are restoring from the backup.
+ "A String",
+ ],
"serviceRevision": { # A managed metastore service that serves metadata queries. # Output only. The revision of the service at the time of backup.
"artifactGcsUri": "A String", # Output only. A Cloud Storage URI (starting with gs://) that specifies where artifacts related to the metastore service are stored.
"createTime": "A String", # Output only. The time when the metastore service was created.
@@ -413,6 +419,9 @@ Method Details
"description": "A String", # The description of the backup.
"endTime": "A String", # Output only. The time when the backup finished creating.
"name": "A String", # Immutable. The relative resource name of the backup, in the following form:projects/{project_number}/locations/{location_id}/services/{service_id}/backups/{backup_id}
+ "restoringServices": [ # Output only. Services that are restoring from the backup.
+ "A String",
+ ],
"serviceRevision": { # A managed metastore service that serves metadata queries. # Output only. The revision of the service at the time of backup.
"artifactGcsUri": "A String", # Output only. A Cloud Storage URI (starting with gs://) that specifies where artifacts related to the metastore service are stored.
"createTime": "A String", # Output only. The time when the metastore service was created.
diff --git a/docs/dyn/metastore_v1alpha.projects.locations.services.databases.html b/docs/dyn/metastore_v1alpha.projects.locations.services.databases.html
new file mode 100644
index 00000000000..885f78f7b6f
--- /dev/null
+++ b/docs/dyn/metastore_v1alpha.projects.locations.services.databases.html
@@ -0,0 +1,263 @@
+
+
+
+Dataproc Metastore API . projects . locations . services . databases
+Instance Methods
+
+ tables()
+
+Returns the tables Resource.
+
+
+ close()
+Close httplib2 connections.
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+Sets the access control policy on the specified resource. Replaces any existing policy.Can return NOT_FOUND, INVALID_ARGUMENT, and PERMISSION_DENIED errors.
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a NOT_FOUND error.Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+ Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being requested. See the operation documentation for the appropriate value for this field. (required)
+ options_requestedPolicyVersion: integer, Optional. The policy format version to be returned.Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected.Requests for policies with any conditional bindings must specify version 3. Policies without any conditional bindings may specify any valid value or leave the field unset.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources.A Policy is a collection of bindings. A binding binds one or more members to a single role. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A role is a named list of permissions; each role can be an IAM predefined role or a user-created custom role.For some types of Google Cloud resources, a binding can also specify a condition, which is a logical expression that allows access to a resource only if the expression evaluates to true. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).JSON example: { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } YAML example: bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the IAM documentation (https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs.If there are AuditConfigs for both allServices and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted.Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, storage.googleapis.com, cloudsql.googleapis.com. allServices is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of members to a role. Optionally, may specify a condition that determines how and when the bindings are applied. Each of the bindings must contain at least one member.
+ { # Associates members with a role.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec.Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding.If the condition evaluates to true, then this binding applies to the current request.If the condition evaluates to false, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. members can have the following values: allUsers: A special identifier that represents anyone who is on the internet; with or without a Google account. allAuthenticatedUsers: A special identifier that represents anyone who is authenticated with a Google account or a service account. user:{emailid}: An email address that represents a specific Google account. For example, alice@example.com . serviceAccount:{emailid}: An email address that represents a service account. For example, my-other-app@appspot.gserviceaccount.com. group:{emailid}: An email address that represents a Google group. For example, admins@example.com. deleted:user:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a user that has been recently deleted. For example, alice@example.com?uid=123456789012345678901. If the user is recovered, this value reverts to user:{emailid} and the recovered user retains the role in the binding. deleted:serviceAccount:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901. If the service account is undeleted, this value reverts to serviceAccount:{emailid} and the undeleted service account retains the role in the binding. deleted:group:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, admins@example.com?uid=123456789012345678901. If the group is recovered, this value reverts to group:{emailid} and the recovered group retains the role in the binding. domain:{domain}: The G Suite domain (primary) that represents all the users of that domain. For example, google.com or example.com.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to members. For example, roles/viewer, roles/editor, or roles/owner.
+ },
+ ],
+ "etag": "A String", # etag is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the etag in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An etag is returned in the response to getIamPolicy, and systems are expected to put that etag in the request to setIamPolicy to ensure that their change will be applied to the same version of the policy.Important: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.
+ "version": 42, # Specifies the format of the policy.Valid values are 0, 1, and 3. Requests that specify an invalid value are rejected.Any operation that affects conditional role bindings must specify version 3. This requirement applies to the following operations: Getting a policy that includes a conditional role binding Adding a conditional role binding to a policy Changing a conditional role binding in a policy Removing any role binding, with or without a condition, from a policy that includes conditionsImportant: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+ Sets the access control policy on the specified resource. Replaces any existing policy.Can return NOT_FOUND, INVALID_ARGUMENT, and PERMISSION_DENIED errors.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being specified. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for SetIamPolicy method.
+ "policy": { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources.A Policy is a collection of bindings. A binding binds one or more members to a single role. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A role is a named list of permissions; each role can be an IAM predefined role or a user-created custom role.For some types of Google Cloud resources, a binding can also specify a condition, which is a logical expression that allows access to a resource only if the expression evaluates to true. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).JSON example: { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } YAML example: bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the IAM documentation (https://cloud.google.com/iam/docs/). # REQUIRED: The complete policy to be applied to the resource. The size of the policy is limited to a few 10s of KB. An empty policy is a valid policy but certain Cloud Platform services (such as Projects) might reject them.
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs.If there are AuditConfigs for both allServices and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted.Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, storage.googleapis.com, cloudsql.googleapis.com. allServices is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of members to a role. Optionally, may specify a condition that determines how and when the bindings are applied. Each of the bindings must contain at least one member.
+ { # Associates members with a role.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec.Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding.If the condition evaluates to true, then this binding applies to the current request.If the condition evaluates to false, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. members can have the following values: allUsers: A special identifier that represents anyone who is on the internet; with or without a Google account. allAuthenticatedUsers: A special identifier that represents anyone who is authenticated with a Google account or a service account. user:{emailid}: An email address that represents a specific Google account. For example, alice@example.com . serviceAccount:{emailid}: An email address that represents a service account. For example, my-other-app@appspot.gserviceaccount.com. group:{emailid}: An email address that represents a Google group. For example, admins@example.com. deleted:user:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a user that has been recently deleted. For example, alice@example.com?uid=123456789012345678901. If the user is recovered, this value reverts to user:{emailid} and the recovered user retains the role in the binding. deleted:serviceAccount:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901. If the service account is undeleted, this value reverts to serviceAccount:{emailid} and the undeleted service account retains the role in the binding. deleted:group:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, admins@example.com?uid=123456789012345678901. If the group is recovered, this value reverts to group:{emailid} and the recovered group retains the role in the binding. domain:{domain}: The G Suite domain (primary) that represents all the users of that domain. For example, google.com or example.com.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to members. For example, roles/viewer, roles/editor, or roles/owner.
+ },
+ ],
+ "etag": "A String", # etag is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the etag in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An etag is returned in the response to getIamPolicy, and systems are expected to put that etag in the request to setIamPolicy to ensure that their change will be applied to the same version of the policy.Important: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.
+ "version": 42, # Specifies the format of the policy.Valid values are 0, 1, and 3. Requests that specify an invalid value are rejected.Any operation that affects conditional role bindings must specify version 3. This requirement applies to the following operations: Getting a policy that includes a conditional role binding Adding a conditional role binding to a policy Changing a conditional role binding in a policy Removing any role binding, with or without a condition, from a policy that includes conditionsImportant: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+ },
+ "updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only the fields in the mask will be modified. If no mask is provided, the following default mask is used:paths: "bindings, etag"
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources.A Policy is a collection of bindings. A binding binds one or more members to a single role. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A role is a named list of permissions; each role can be an IAM predefined role or a user-created custom role.For some types of Google Cloud resources, a binding can also specify a condition, which is a logical expression that allows access to a resource only if the expression evaluates to true. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).JSON example: { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } YAML example: bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the IAM documentation (https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs.If there are AuditConfigs for both allServices and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted.Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, storage.googleapis.com, cloudsql.googleapis.com. allServices is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of members to a role. Optionally, may specify a condition that determines how and when the bindings are applied. Each of the bindings must contain at least one member.
+ { # Associates members with a role.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec.Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding.If the condition evaluates to true, then this binding applies to the current request.If the condition evaluates to false, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. members can have the following values: allUsers: A special identifier that represents anyone who is on the internet; with or without a Google account. allAuthenticatedUsers: A special identifier that represents anyone who is authenticated with a Google account or a service account. user:{emailid}: An email address that represents a specific Google account. For example, alice@example.com . serviceAccount:{emailid}: An email address that represents a service account. For example, my-other-app@appspot.gserviceaccount.com. group:{emailid}: An email address that represents a Google group. For example, admins@example.com. deleted:user:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a user that has been recently deleted. For example, alice@example.com?uid=123456789012345678901. If the user is recovered, this value reverts to user:{emailid} and the recovered user retains the role in the binding. deleted:serviceAccount:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901. If the service account is undeleted, this value reverts to serviceAccount:{emailid} and the undeleted service account retains the role in the binding. deleted:group:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, admins@example.com?uid=123456789012345678901. If the group is recovered, this value reverts to group:{emailid} and the recovered group retains the role in the binding. domain:{domain}: The G Suite domain (primary) that represents all the users of that domain. For example, google.com or example.com.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to members. For example, roles/viewer, roles/editor, or roles/owner.
+ },
+ ],
+ "etag": "A String", # etag is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the etag in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An etag is returned in the response to getIamPolicy, and systems are expected to put that etag in the request to setIamPolicy to ensure that their change will be applied to the same version of the policy.Important: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.
+ "version": 42, # Specifies the format of the policy.Valid values are 0, 1, and 3. Requests that specify an invalid value are rejected.Any operation that affects conditional role bindings must specify version 3. This requirement applies to the following operations: Getting a policy that includes a conditional role binding Adding a conditional role binding to a policy Changing a conditional role binding in a policy Removing any role binding, with or without a condition, from a policy that includes conditionsImportant: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+ Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a NOT_FOUND error.Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy detail is being requested. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for TestIamPermissions method.
+ "permissions": [ # The set of permissions to check for the resource. Permissions with wildcards (such as '*' or 'storage.*') are not allowed. For more information see IAM Overview (https://cloud.google.com/iam/docs/overview#permissions).
+ "A String",
+ ],
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response message for TestIamPermissions method.
+ "permissions": [ # A subset of TestPermissionsRequest.permissions that the caller is allowed.
+ "A String",
+ ],
+}
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/metastore_v1alpha.projects.locations.services.databases.tables.html b/docs/dyn/metastore_v1alpha.projects.locations.services.databases.tables.html
new file mode 100644
index 00000000000..22a2cdc166e
--- /dev/null
+++ b/docs/dyn/metastore_v1alpha.projects.locations.services.databases.tables.html
@@ -0,0 +1,258 @@
+
+
+
+Dataproc Metastore API . projects . locations . services . databases . tables
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+Sets the access control policy on the specified resource. Replaces any existing policy.Can return NOT_FOUND, INVALID_ARGUMENT, and PERMISSION_DENIED errors.
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a NOT_FOUND error.Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+ Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being requested. See the operation documentation for the appropriate value for this field. (required)
+ options_requestedPolicyVersion: integer, Optional. The policy format version to be returned.Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected.Requests for policies with any conditional bindings must specify version 3. Policies without any conditional bindings may specify any valid value or leave the field unset.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources.A Policy is a collection of bindings. A binding binds one or more members to a single role. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A role is a named list of permissions; each role can be an IAM predefined role or a user-created custom role.For some types of Google Cloud resources, a binding can also specify a condition, which is a logical expression that allows access to a resource only if the expression evaluates to true. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).JSON example: { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } YAML example: bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the IAM documentation (https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs.If there are AuditConfigs for both allServices and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted.Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, storage.googleapis.com, cloudsql.googleapis.com. allServices is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of members to a role. Optionally, may specify a condition that determines how and when the bindings are applied. Each of the bindings must contain at least one member.
+ { # Associates members with a role.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec.Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding.If the condition evaluates to true, then this binding applies to the current request.If the condition evaluates to false, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. members can have the following values: allUsers: A special identifier that represents anyone who is on the internet; with or without a Google account. allAuthenticatedUsers: A special identifier that represents anyone who is authenticated with a Google account or a service account. user:{emailid}: An email address that represents a specific Google account. For example, alice@example.com . serviceAccount:{emailid}: An email address that represents a service account. For example, my-other-app@appspot.gserviceaccount.com. group:{emailid}: An email address that represents a Google group. For example, admins@example.com. deleted:user:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a user that has been recently deleted. For example, alice@example.com?uid=123456789012345678901. If the user is recovered, this value reverts to user:{emailid} and the recovered user retains the role in the binding. deleted:serviceAccount:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901. If the service account is undeleted, this value reverts to serviceAccount:{emailid} and the undeleted service account retains the role in the binding. deleted:group:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, admins@example.com?uid=123456789012345678901. If the group is recovered, this value reverts to group:{emailid} and the recovered group retains the role in the binding. domain:{domain}: The G Suite domain (primary) that represents all the users of that domain. For example, google.com or example.com.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to members. For example, roles/viewer, roles/editor, or roles/owner.
+ },
+ ],
+ "etag": "A String", # etag is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the etag in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An etag is returned in the response to getIamPolicy, and systems are expected to put that etag in the request to setIamPolicy to ensure that their change will be applied to the same version of the policy.Important: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.
+ "version": 42, # Specifies the format of the policy.Valid values are 0, 1, and 3. Requests that specify an invalid value are rejected.Any operation that affects conditional role bindings must specify version 3. This requirement applies to the following operations: Getting a policy that includes a conditional role binding Adding a conditional role binding to a policy Changing a conditional role binding in a policy Removing any role binding, with or without a condition, from a policy that includes conditionsImportant: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+ Sets the access control policy on the specified resource. Replaces any existing policy.Can return NOT_FOUND, INVALID_ARGUMENT, and PERMISSION_DENIED errors.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being specified. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for SetIamPolicy method.
+ "policy": { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources.A Policy is a collection of bindings. A binding binds one or more members to a single role. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A role is a named list of permissions; each role can be an IAM predefined role or a user-created custom role.For some types of Google Cloud resources, a binding can also specify a condition, which is a logical expression that allows access to a resource only if the expression evaluates to true. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).JSON example: { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } YAML example: bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the IAM documentation (https://cloud.google.com/iam/docs/). # REQUIRED: The complete policy to be applied to the resource. The size of the policy is limited to a few 10s of KB. An empty policy is a valid policy but certain Cloud Platform services (such as Projects) might reject them.
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs.If there are AuditConfigs for both allServices and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted.Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, storage.googleapis.com, cloudsql.googleapis.com. allServices is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of members to a role. Optionally, may specify a condition that determines how and when the bindings are applied. Each of the bindings must contain at least one member.
+ { # Associates members with a role.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec.Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding.If the condition evaluates to true, then this binding applies to the current request.If the condition evaluates to false, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. members can have the following values: allUsers: A special identifier that represents anyone who is on the internet; with or without a Google account. allAuthenticatedUsers: A special identifier that represents anyone who is authenticated with a Google account or a service account. user:{emailid}: An email address that represents a specific Google account. For example, alice@example.com . serviceAccount:{emailid}: An email address that represents a service account. For example, my-other-app@appspot.gserviceaccount.com. group:{emailid}: An email address that represents a Google group. For example, admins@example.com. deleted:user:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a user that has been recently deleted. For example, alice@example.com?uid=123456789012345678901. If the user is recovered, this value reverts to user:{emailid} and the recovered user retains the role in the binding. deleted:serviceAccount:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901. If the service account is undeleted, this value reverts to serviceAccount:{emailid} and the undeleted service account retains the role in the binding. deleted:group:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, admins@example.com?uid=123456789012345678901. If the group is recovered, this value reverts to group:{emailid} and the recovered group retains the role in the binding. domain:{domain}: The G Suite domain (primary) that represents all the users of that domain. For example, google.com or example.com.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to members. For example, roles/viewer, roles/editor, or roles/owner.
+ },
+ ],
+ "etag": "A String", # etag is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the etag in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An etag is returned in the response to getIamPolicy, and systems are expected to put that etag in the request to setIamPolicy to ensure that their change will be applied to the same version of the policy.Important: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.
+ "version": 42, # Specifies the format of the policy.Valid values are 0, 1, and 3. Requests that specify an invalid value are rejected.Any operation that affects conditional role bindings must specify version 3. This requirement applies to the following operations: Getting a policy that includes a conditional role binding Adding a conditional role binding to a policy Changing a conditional role binding in a policy Removing any role binding, with or without a condition, from a policy that includes conditionsImportant: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+ },
+ "updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only the fields in the mask will be modified. If no mask is provided, the following default mask is used:paths: "bindings, etag"
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources.A Policy is a collection of bindings. A binding binds one or more members to a single role. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A role is a named list of permissions; each role can be an IAM predefined role or a user-created custom role.For some types of Google Cloud resources, a binding can also specify a condition, which is a logical expression that allows access to a resource only if the expression evaluates to true. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).JSON example: { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } YAML example: bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the IAM documentation (https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs.If there are AuditConfigs for both allServices and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted.Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, storage.googleapis.com, cloudsql.googleapis.com. allServices is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of members to a role. Optionally, may specify a condition that determines how and when the bindings are applied. Each of the bindings must contain at least one member.
+ { # Associates members with a role.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec.Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding.If the condition evaluates to true, then this binding applies to the current request.If the condition evaluates to false, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. members can have the following values: allUsers: A special identifier that represents anyone who is on the internet; with or without a Google account. allAuthenticatedUsers: A special identifier that represents anyone who is authenticated with a Google account or a service account. user:{emailid}: An email address that represents a specific Google account. For example, alice@example.com . serviceAccount:{emailid}: An email address that represents a service account. For example, my-other-app@appspot.gserviceaccount.com. group:{emailid}: An email address that represents a Google group. For example, admins@example.com. deleted:user:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a user that has been recently deleted. For example, alice@example.com?uid=123456789012345678901. If the user is recovered, this value reverts to user:{emailid} and the recovered user retains the role in the binding. deleted:serviceAccount:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901. If the service account is undeleted, this value reverts to serviceAccount:{emailid} and the undeleted service account retains the role in the binding. deleted:group:{emailid}?uid={uniqueid}: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, admins@example.com?uid=123456789012345678901. If the group is recovered, this value reverts to group:{emailid} and the recovered group retains the role in the binding. domain:{domain}: The G Suite domain (primary) that represents all the users of that domain. For example, google.com or example.com.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to members. For example, roles/viewer, roles/editor, or roles/owner.
+ },
+ ],
+ "etag": "A String", # etag is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the etag in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An etag is returned in the response to getIamPolicy, and systems are expected to put that etag in the request to setIamPolicy to ensure that their change will be applied to the same version of the policy.Important: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.
+ "version": 42, # Specifies the format of the policy.Valid values are 0, 1, and 3. Requests that specify an invalid value are rejected.Any operation that affects conditional role bindings must specify version 3. This requirement applies to the following operations: Getting a policy that includes a conditional role binding Adding a conditional role binding to a policy Changing a conditional role binding in a policy Removing any role binding, with or without a condition, from a policy that includes conditionsImportant: If you use IAM Conditions, you must include the etag field whenever you call setIamPolicy. If you omit this field, then IAM allows you to overwrite a version 3 policy with a version 1 policy, and all of the conditions in the version 3 policy are lost.If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset.To learn which resources support conditions in their IAM policies, see the IAM documentation (https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+ Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a NOT_FOUND error.Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy detail is being requested. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for TestIamPermissions method.
+ "permissions": [ # The set of permissions to check for the resource. Permissions with wildcards (such as '*' or 'storage.*') are not allowed. For more information see IAM Overview (https://cloud.google.com/iam/docs/overview#permissions).
+ "A String",
+ ],
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response message for TestIamPermissions method.
+ "permissions": [ # A subset of TestPermissionsRequest.permissions that the caller is allowed.
+ "A String",
+ ],
+}
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/metastore_v1alpha.projects.locations.services.html b/docs/dyn/metastore_v1alpha.projects.locations.services.html
index 47018bb92fe..b0eecccf27b 100644
--- a/docs/dyn/metastore_v1alpha.projects.locations.services.html
+++ b/docs/dyn/metastore_v1alpha.projects.locations.services.html
@@ -79,6 +79,11 @@ Instance Methods
Returns the backups Resource.
+
+ databases()
+
+Returns the databases Resource.
+
diff --git a/docs/dyn/metastore_v1beta.projects.locations.services.backups.html b/docs/dyn/metastore_v1beta.projects.locations.services.backups.html
index 6cae1f472bf..97767ca3c30 100644
--- a/docs/dyn/metastore_v1beta.projects.locations.services.backups.html
+++ b/docs/dyn/metastore_v1beta.projects.locations.services.backups.html
@@ -121,6 +121,9 @@ Method Details
"description": "A String", # The description of the backup.
"endTime": "A String", # Output only. The time when the backup finished creating.
"name": "A String", # Immutable. The relative resource name of the backup, in the following form:projects/{project_number}/locations/{location_id}/services/{service_id}/backups/{backup_id}
+ "restoringServices": [ # Output only. Services that are restoring from the backup.
+ "A String",
+ ],
"serviceRevision": { # A managed metastore service that serves metadata queries. # Output only. The revision of the service at the time of backup.
"artifactGcsUri": "A String", # Output only. A Cloud Storage URI (starting with gs://) that specifies where artifacts related to the metastore service are stored.
"createTime": "A String", # Output only. The time when the metastore service was created.
@@ -273,6 +276,9 @@ Method Details
"description": "A String", # The description of the backup.
"endTime": "A String", # Output only. The time when the backup finished creating.
"name": "A String", # Immutable. The relative resource name of the backup, in the following form:projects/{project_number}/locations/{location_id}/services/{service_id}/backups/{backup_id}
+ "restoringServices": [ # Output only. Services that are restoring from the backup.
+ "A String",
+ ],
"serviceRevision": { # A managed metastore service that serves metadata queries. # Output only. The revision of the service at the time of backup.
"artifactGcsUri": "A String", # Output only. A Cloud Storage URI (starting with gs://) that specifies where artifacts related to the metastore service are stored.
"createTime": "A String", # Output only. The time when the metastore service was created.
@@ -413,6 +419,9 @@ Method Details
"description": "A String", # The description of the backup.
"endTime": "A String", # Output only. The time when the backup finished creating.
"name": "A String", # Immutable. The relative resource name of the backup, in the following form:projects/{project_number}/locations/{location_id}/services/{service_id}/backups/{backup_id}
+ "restoringServices": [ # Output only. Services that are restoring from the backup.
+ "A String",
+ ],
"serviceRevision": { # A managed metastore service that serves metadata queries. # Output only. The revision of the service at the time of backup.
"artifactGcsUri": "A String", # Output only. A Cloud Storage URI (starting with gs://) that specifies where artifacts related to the metastore service are stored.
"createTime": "A String", # Output only. The time when the metastore service was created.
diff --git a/docs/dyn/ml_v1.projects.jobs.html b/docs/dyn/ml_v1.projects.jobs.html
index 243645d9fd1..4a1387ae469 100644
--- a/docs/dyn/ml_v1.projects.jobs.html
+++ b/docs/dyn/ml_v1.projects.jobs.html
@@ -180,7 +180,7 @@ Method Details
"args": [ # Optional. Command-line arguments passed to the training application when it starts. If your job uses a custom container, then the arguments are passed to the container's `ENTRYPOINT` command.
"A String",
],
- "enableWebAccess": True or False, # Optional. Whether to enable web access for the training job.
+ "enableWebAccess": True or False, # Optional. Whether you want AI Platform Training to enable [interactive shell access](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) to training containers. If set to `true`, you can access interactive shells at the URIs given by TrainingOutput.web_access_uris or HyperparameterOutput.web_access_uris (within TrainingOutput.trials).
"encryptionConfig": { # Represents a custom encryption key configuration that can be applied to a resource. # Optional. Options for using customer-managed encryption keys (CMEK) to protect resources created by a training job, instead of using Google's default encryption. If this is set, then all resources created by the training job will be encrypted with the customer-managed encryption key that you specify. [Learn how and when to use CMEK with AI Platform Training](/ai-platform/training/docs/cmek).
"kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key used to protect a resource, such as a training job. It has the following format: `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
},
@@ -344,12 +344,12 @@ Method Details
"startTime": "A String", # Output only. Start time for the trial.
"state": "A String", # Output only. The detailed state of the trial.
"trialId": "A String", # The trial id for these results.
- "webAccessUris": { # The web URIs for the training job. Currently for debug terminal access to the job. Only set for in-progress hyperparameter tuning trials with web access enabled.
+ "webAccessUris": { # URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if this trial is part of a hyperparameter tuning job and the job's training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
],
- "webAccessUris": { # Output only. The web URIs for the training job. Currently for debug terminal access to the job.
+ "webAccessUris": { # Output only. URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
@@ -400,7 +400,7 @@ Method Details
"args": [ # Optional. Command-line arguments passed to the training application when it starts. If your job uses a custom container, then the arguments are passed to the container's `ENTRYPOINT` command.
"A String",
],
- "enableWebAccess": True or False, # Optional. Whether to enable web access for the training job.
+ "enableWebAccess": True or False, # Optional. Whether you want AI Platform Training to enable [interactive shell access](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) to training containers. If set to `true`, you can access interactive shells at the URIs given by TrainingOutput.web_access_uris or HyperparameterOutput.web_access_uris (within TrainingOutput.trials).
"encryptionConfig": { # Represents a custom encryption key configuration that can be applied to a resource. # Optional. Options for using customer-managed encryption keys (CMEK) to protect resources created by a training job, instead of using Google's default encryption. If this is set, then all resources created by the training job will be encrypted with the customer-managed encryption key that you specify. [Learn how and when to use CMEK with AI Platform Training](/ai-platform/training/docs/cmek).
"kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key used to protect a resource, such as a training job. It has the following format: `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
},
@@ -564,12 +564,12 @@ Method Details
"startTime": "A String", # Output only. Start time for the trial.
"state": "A String", # Output only. The detailed state of the trial.
"trialId": "A String", # The trial id for these results.
- "webAccessUris": { # The web URIs for the training job. Currently for debug terminal access to the job. Only set for in-progress hyperparameter tuning trials with web access enabled.
+ "webAccessUris": { # URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if this trial is part of a hyperparameter tuning job and the job's training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
],
- "webAccessUris": { # Output only. The web URIs for the training job. Currently for debug terminal access to the job.
+ "webAccessUris": { # Output only. URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
@@ -627,7 +627,7 @@ Method Details
"args": [ # Optional. Command-line arguments passed to the training application when it starts. If your job uses a custom container, then the arguments are passed to the container's `ENTRYPOINT` command.
"A String",
],
- "enableWebAccess": True or False, # Optional. Whether to enable web access for the training job.
+ "enableWebAccess": True or False, # Optional. Whether you want AI Platform Training to enable [interactive shell access](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) to training containers. If set to `true`, you can access interactive shells at the URIs given by TrainingOutput.web_access_uris or HyperparameterOutput.web_access_uris (within TrainingOutput.trials).
"encryptionConfig": { # Represents a custom encryption key configuration that can be applied to a resource. # Optional. Options for using customer-managed encryption keys (CMEK) to protect resources created by a training job, instead of using Google's default encryption. If this is set, then all resources created by the training job will be encrypted with the customer-managed encryption key that you specify. [Learn how and when to use CMEK with AI Platform Training](/ai-platform/training/docs/cmek).
"kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key used to protect a resource, such as a training job. It has the following format: `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
},
@@ -791,12 +791,12 @@ Method Details
"startTime": "A String", # Output only. Start time for the trial.
"state": "A String", # Output only. The detailed state of the trial.
"trialId": "A String", # The trial id for these results.
- "webAccessUris": { # The web URIs for the training job. Currently for debug terminal access to the job. Only set for in-progress hyperparameter tuning trials with web access enabled.
+ "webAccessUris": { # URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if this trial is part of a hyperparameter tuning job and the job's training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
],
- "webAccessUris": { # Output only. The web URIs for the training job. Currently for debug terminal access to the job.
+ "webAccessUris": { # Output only. URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
@@ -907,7 +907,7 @@ Method Details
"args": [ # Optional. Command-line arguments passed to the training application when it starts. If your job uses a custom container, then the arguments are passed to the container's `ENTRYPOINT` command.
"A String",
],
- "enableWebAccess": True or False, # Optional. Whether to enable web access for the training job.
+ "enableWebAccess": True or False, # Optional. Whether you want AI Platform Training to enable [interactive shell access](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) to training containers. If set to `true`, you can access interactive shells at the URIs given by TrainingOutput.web_access_uris or HyperparameterOutput.web_access_uris (within TrainingOutput.trials).
"encryptionConfig": { # Represents a custom encryption key configuration that can be applied to a resource. # Optional. Options for using customer-managed encryption keys (CMEK) to protect resources created by a training job, instead of using Google's default encryption. If this is set, then all resources created by the training job will be encrypted with the customer-managed encryption key that you specify. [Learn how and when to use CMEK with AI Platform Training](/ai-platform/training/docs/cmek).
"kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key used to protect a resource, such as a training job. It has the following format: `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
},
@@ -1071,12 +1071,12 @@ Method Details
"startTime": "A String", # Output only. Start time for the trial.
"state": "A String", # Output only. The detailed state of the trial.
"trialId": "A String", # The trial id for these results.
- "webAccessUris": { # The web URIs for the training job. Currently for debug terminal access to the job. Only set for in-progress hyperparameter tuning trials with web access enabled.
+ "webAccessUris": { # URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if this trial is part of a hyperparameter tuning job and the job's training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
],
- "webAccessUris": { # Output only. The web URIs for the training job. Currently for debug terminal access to the job.
+ "webAccessUris": { # Output only. URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
@@ -1146,7 +1146,7 @@ Method Details
"args": [ # Optional. Command-line arguments passed to the training application when it starts. If your job uses a custom container, then the arguments are passed to the container's `ENTRYPOINT` command.
"A String",
],
- "enableWebAccess": True or False, # Optional. Whether to enable web access for the training job.
+ "enableWebAccess": True or False, # Optional. Whether you want AI Platform Training to enable [interactive shell access](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) to training containers. If set to `true`, you can access interactive shells at the URIs given by TrainingOutput.web_access_uris or HyperparameterOutput.web_access_uris (within TrainingOutput.trials).
"encryptionConfig": { # Represents a custom encryption key configuration that can be applied to a resource. # Optional. Options for using customer-managed encryption keys (CMEK) to protect resources created by a training job, instead of using Google's default encryption. If this is set, then all resources created by the training job will be encrypted with the customer-managed encryption key that you specify. [Learn how and when to use CMEK with AI Platform Training](/ai-platform/training/docs/cmek).
"kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key used to protect a resource, such as a training job. It has the following format: `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
},
@@ -1310,12 +1310,12 @@ Method Details
"startTime": "A String", # Output only. Start time for the trial.
"state": "A String", # Output only. The detailed state of the trial.
"trialId": "A String", # The trial id for these results.
- "webAccessUris": { # The web URIs for the training job. Currently for debug terminal access to the job. Only set for in-progress hyperparameter tuning trials with web access enabled.
+ "webAccessUris": { # URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if this trial is part of a hyperparameter tuning job and the job's training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
],
- "webAccessUris": { # Output only. The web URIs for the training job. Currently for debug terminal access to the job.
+ "webAccessUris": { # Output only. URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
@@ -1367,7 +1367,7 @@ Method Details
"args": [ # Optional. Command-line arguments passed to the training application when it starts. If your job uses a custom container, then the arguments are passed to the container's `ENTRYPOINT` command.
"A String",
],
- "enableWebAccess": True or False, # Optional. Whether to enable web access for the training job.
+ "enableWebAccess": True or False, # Optional. Whether you want AI Platform Training to enable [interactive shell access](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) to training containers. If set to `true`, you can access interactive shells at the URIs given by TrainingOutput.web_access_uris or HyperparameterOutput.web_access_uris (within TrainingOutput.trials).
"encryptionConfig": { # Represents a custom encryption key configuration that can be applied to a resource. # Optional. Options for using customer-managed encryption keys (CMEK) to protect resources created by a training job, instead of using Google's default encryption. If this is set, then all resources created by the training job will be encrypted with the customer-managed encryption key that you specify. [Learn how and when to use CMEK with AI Platform Training](/ai-platform/training/docs/cmek).
"kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key used to protect a resource, such as a training job. It has the following format: `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
},
@@ -1531,12 +1531,12 @@ Method Details
"startTime": "A String", # Output only. Start time for the trial.
"state": "A String", # Output only. The detailed state of the trial.
"trialId": "A String", # The trial id for these results.
- "webAccessUris": { # The web URIs for the training job. Currently for debug terminal access to the job. Only set for in-progress hyperparameter tuning trials with web access enabled.
+ "webAccessUris": { # URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if this trial is part of a hyperparameter tuning job and the job's training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
],
- "webAccessUris": { # Output only. The web URIs for the training job. Currently for debug terminal access to the job.
+ "webAccessUris": { # Output only. URIs for accessing [interactive shells](https://cloud.google.com/ai-platform/training/docs/monitor-debug-interactive-shell) (one URI for each training node). Only available if training_input.enable_web_access is `true`. The keys are names of each node in the training job; for example, `master-replica-0` for the master node, `worker-replica-0` for the first worker, and `ps-replica-0` for the first parameter server. The values are the URIs for each node's interactive shell.
"a_key": "A String",
},
},
diff --git a/docs/dyn/monitoring_v1.locations.global_.metricsScopes.html b/docs/dyn/monitoring_v1.locations.global_.metricsScopes.html
index 8c35e5c7de4..cb835b4ccc3 100644
--- a/docs/dyn/monitoring_v1.locations.global_.metricsScopes.html
+++ b/docs/dyn/monitoring_v1.locations.global_.metricsScopes.html
@@ -86,7 +86,7 @@ Instance Methods
get(name, x__xgafv=None)
Returns a specific Metrics Scope.
- listMetricScopesByMonitoredProject(monitoredResourceContainer=None, x__xgafv=None)
+ listMetricsScopesByMonitoredProject(monitoredResourceContainer=None, x__xgafv=None)
Returns a list of every Metrics Scope that a specific MonitoredProject has been added to. The metrics scope representing the specified monitored project will always be the first entry in the response.
Method Details
@@ -122,7 +122,7 @@ Method Details
- listMetricScopesByMonitoredProject(monitoredResourceContainer=None, x__xgafv=None)
+ listMetricsScopesByMonitoredProject(monitoredResourceContainer=None, x__xgafv=None)
Returns a list of every Metrics Scope that a specific MonitoredProject has been added to. The metrics scope representing the specified monitored project will always be the first entry in the response.
Args:
diff --git a/docs/dyn/monitoring_v3.projects.monitoredResourceDescriptors.html b/docs/dyn/monitoring_v3.projects.monitoredResourceDescriptors.html
index b3ef2338950..ef14f48bd79 100644
--- a/docs/dyn/monitoring_v3.projects.monitoredResourceDescriptors.html
+++ b/docs/dyn/monitoring_v3.projects.monitoredResourceDescriptors.html
@@ -118,7 +118,7 @@ Method Details
],
"launchStage": "A String", # Optional. The launch stage of the monitored resource definition.
"name": "A String", # Optional. The resource name of the monitored resource descriptor: "projects/{project_id}/monitoredResourceDescriptors/{type}" where {type} is the value of the type field in this object and {project_id} is a project ID that provides API-specific context for accessing the type. APIs that do not use project information can use the resource name format "monitoredResourceDescriptors/{type}".
- "type": "A String", # Required. The monitored resource type. For example, the type "cloudsql_database" represents databases in Google Cloud SQL.
+ "type": "A String", # Required. The monitored resource type. For example, the type "cloudsql_database" represents databases in Google Cloud SQL. For a list of types, see Monitoring resource types (https://cloud.google.com/monitoring/api/resources) and Logging resource types (https://cloud.google.com/logging/docs/api/v2/resource-list).
}
@@ -154,7 +154,7 @@ Method Details
],
"launchStage": "A String", # Optional. The launch stage of the monitored resource definition.
"name": "A String", # Optional. The resource name of the monitored resource descriptor: "projects/{project_id}/monitoredResourceDescriptors/{type}" where {type} is the value of the type field in this object and {project_id} is a project ID that provides API-specific context for accessing the type. APIs that do not use project information can use the resource name format "monitoredResourceDescriptors/{type}".
- "type": "A String", # Required. The monitored resource type. For example, the type "cloudsql_database" represents databases in Google Cloud SQL.
+ "type": "A String", # Required. The monitored resource type. For example, the type "cloudsql_database" represents databases in Google Cloud SQL. For a list of types, see Monitoring resource types (https://cloud.google.com/monitoring/api/resources) and Logging resource types (https://cloud.google.com/logging/docs/api/v2/resource-list).
},
],
}
diff --git a/docs/dyn/monitoring_v3.projects.uptimeCheckConfigs.html b/docs/dyn/monitoring_v3.projects.uptimeCheckConfigs.html
index 5d26be855ba..59507ad4cbb 100644
--- a/docs/dyn/monitoring_v3.projects.uptimeCheckConfigs.html
+++ b/docs/dyn/monitoring_v3.projects.uptimeCheckConfigs.html
@@ -146,7 +146,7 @@ Method Details
},
],
"isInternal": True or False, # If this is true, then checks are made only from the 'internal_checkers'. If it is false, then checks are made only from the 'selected_regions'. It is an error to provide 'selected_regions' when is_internal is true, or to provide 'internal_checkers' when is_internal is false.
- "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer
+ "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer k8s_service
"labels": { # Required. Values for all of the labels listed in the associated monitored resource descriptor. For example, Compute Engine VM instances use the labels "project_id", "instance_id", and "zone".
"a_key": "A String",
},
@@ -211,7 +211,7 @@ Method Details
},
],
"isInternal": True or False, # If this is true, then checks are made only from the 'internal_checkers'. If it is false, then checks are made only from the 'selected_regions'. It is an error to provide 'selected_regions' when is_internal is true, or to provide 'internal_checkers' when is_internal is false.
- "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer
+ "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer k8s_service
"labels": { # Required. Values for all of the labels listed in the associated monitored resource descriptor. For example, Compute Engine VM instances use the labels "project_id", "instance_id", and "zone".
"a_key": "A String",
},
@@ -301,7 +301,7 @@ Method Details
},
],
"isInternal": True or False, # If this is true, then checks are made only from the 'internal_checkers'. If it is false, then checks are made only from the 'selected_regions'. It is an error to provide 'selected_regions' when is_internal is true, or to provide 'internal_checkers' when is_internal is false.
- "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer
+ "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer k8s_service
"labels": { # Required. Values for all of the labels listed in the associated monitored resource descriptor. For example, Compute Engine VM instances use the labels "project_id", "instance_id", and "zone".
"a_key": "A String",
},
@@ -379,7 +379,7 @@ Method Details
},
],
"isInternal": True or False, # If this is true, then checks are made only from the 'internal_checkers'. If it is false, then checks are made only from the 'selected_regions'. It is an error to provide 'selected_regions' when is_internal is true, or to provide 'internal_checkers' when is_internal is false.
- "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer
+ "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer k8s_service
"labels": { # Required. Values for all of the labels listed in the associated monitored resource descriptor. For example, Compute Engine VM instances use the labels "project_id", "instance_id", and "zone".
"a_key": "A String",
},
@@ -462,7 +462,7 @@ Method Details
},
],
"isInternal": True or False, # If this is true, then checks are made only from the 'internal_checkers'. If it is false, then checks are made only from the 'selected_regions'. It is an error to provide 'selected_regions' when is_internal is true, or to provide 'internal_checkers' when is_internal is false.
- "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer
+ "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer k8s_service
"labels": { # Required. Values for all of the labels listed in the associated monitored resource descriptor. For example, Compute Engine VM instances use the labels "project_id", "instance_id", and "zone".
"a_key": "A String",
},
@@ -528,7 +528,7 @@ Method Details
},
],
"isInternal": True or False, # If this is true, then checks are made only from the 'internal_checkers'. If it is false, then checks are made only from the 'selected_regions'. It is an error to provide 'selected_regions' when is_internal is true, or to provide 'internal_checkers' when is_internal is false.
- "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer
+ "monitoredResource": { # An object representing a resource that can be used for monitoring, logging, billing, or other purposes. Examples include virtual machine instances, databases, and storage devices such as disks. The type field identifies a MonitoredResourceDescriptor object that describes the resource's schema. Information in the labels field identifies the actual resource and its attributes according to the schema. For example, a particular Compute Engine VM instance could be represented by the following object, because the MonitoredResourceDescriptor for "gce_instance" has labels "instance_id" and "zone": { "type": "gce_instance", "labels": { "instance_id": "12345678901234", "zone": "us-central1-a" }} # The monitored resource (https://cloud.google.com/monitoring/api/resources) associated with the configuration. The following monitored resource types are valid for this field: uptime_url, gce_instance, gae_app, aws_ec2_instance, aws_elb_load_balancer k8s_service
"labels": { # Required. Values for all of the labels listed in the associated monitored resource descriptor. For example, Compute Engine VM instances use the labels "project_id", "instance_id", and "zone".
"a_key": "A String",
},
diff --git a/docs/dyn/networkservices_v1.html b/docs/dyn/networkservices_v1.html
new file mode 100644
index 00000000000..7f0d1324329
--- /dev/null
+++ b/docs/dyn/networkservices_v1.html
@@ -0,0 +1,111 @@
+
+
+
+Network Services API
+Instance Methods
+
+ projects()
+
+Returns the projects Resource.
+
+
+ close()
+Close httplib2 connections.
+
+Create a BatchHttpRequest object based on the discovery document.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ new_batch_http_request()
+ Create a BatchHttpRequest object based on the discovery document.
+
+ Args:
+ callback: callable, A callback to be called for each response, of the
+ form callback(id, response, exception). The first parameter is the
+ request id, and the second is the deserialized response object. The
+ third is an apiclient.errors.HttpError exception object if an HTTP
+ error occurred while processing the request, or None if no error
+ occurred.
+
+ Returns:
+ A BatchHttpRequest object based on the discovery document.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/networkservices_v1.projects.html b/docs/dyn/networkservices_v1.projects.html
new file mode 100644
index 00000000000..dea4eedac38
--- /dev/null
+++ b/docs/dyn/networkservices_v1.projects.html
@@ -0,0 +1,91 @@
+
+
+
+Network Services API . projects
+Instance Methods
+
+ locations()
+
+Returns the locations Resource.
+
+
+ close()
+Close httplib2 connections.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/networkservices_v1.projects.locations.edgeCacheKeysets.html b/docs/dyn/networkservices_v1.projects.locations.edgeCacheKeysets.html
new file mode 100644
index 00000000000..c09e59fb1ba
--- /dev/null
+++ b/docs/dyn/networkservices_v1.projects.locations.edgeCacheKeysets.html
@@ -0,0 +1,258 @@
+
+
+
+Network Services API . projects . locations . edgeCacheKeysets
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+ Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being requested. See the operation documentation for the appropriate value for this field. (required)
+ options_requestedPolicyVersion: integer, Optional. The policy format version to be returned. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional bindings must specify version 3. Policies without any conditional bindings may specify any valid value or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+ Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being specified. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for `SetIamPolicy` method.
+ "policy": { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/). # REQUIRED: The complete policy to be applied to the `resource`. The size of the policy is limited to a few 10s of KB. An empty policy is a valid policy but certain Cloud Platform services (such as Projects) might reject them.
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ },
+ "updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only the fields in the mask will be modified. If no mask is provided, the following default mask is used: `paths: "bindings, etag"`
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+ Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy detail is being requested. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for `TestIamPermissions` method.
+ "permissions": [ # The set of permissions to check for the `resource`. Permissions with wildcards (such as '*' or 'storage.*') are not allowed. For more information see [IAM Overview](https://cloud.google.com/iam/docs/overview#permissions).
+ "A String",
+ ],
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response message for `TestIamPermissions` method.
+ "permissions": [ # A subset of `TestPermissionsRequest.permissions` that the caller is allowed.
+ "A String",
+ ],
+}
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/networkservices_v1.projects.locations.edgeCacheOrigins.html b/docs/dyn/networkservices_v1.projects.locations.edgeCacheOrigins.html
new file mode 100644
index 00000000000..27c9c5781c3
--- /dev/null
+++ b/docs/dyn/networkservices_v1.projects.locations.edgeCacheOrigins.html
@@ -0,0 +1,258 @@
+
+
+
+Network Services API . projects . locations . edgeCacheOrigins
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+ Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being requested. See the operation documentation for the appropriate value for this field. (required)
+ options_requestedPolicyVersion: integer, Optional. The policy format version to be returned. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional bindings must specify version 3. Policies without any conditional bindings may specify any valid value or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+ Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being specified. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for `SetIamPolicy` method.
+ "policy": { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/). # REQUIRED: The complete policy to be applied to the `resource`. The size of the policy is limited to a few 10s of KB. An empty policy is a valid policy but certain Cloud Platform services (such as Projects) might reject them.
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ },
+ "updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only the fields in the mask will be modified. If no mask is provided, the following default mask is used: `paths: "bindings, etag"`
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+ Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy detail is being requested. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for `TestIamPermissions` method.
+ "permissions": [ # The set of permissions to check for the `resource`. Permissions with wildcards (such as '*' or 'storage.*') are not allowed. For more information see [IAM Overview](https://cloud.google.com/iam/docs/overview#permissions).
+ "A String",
+ ],
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response message for `TestIamPermissions` method.
+ "permissions": [ # A subset of `TestPermissionsRequest.permissions` that the caller is allowed.
+ "A String",
+ ],
+}
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/networkservices_v1.projects.locations.edgeCacheServices.html b/docs/dyn/networkservices_v1.projects.locations.edgeCacheServices.html
new file mode 100644
index 00000000000..b6d86f11a2d
--- /dev/null
+++ b/docs/dyn/networkservices_v1.projects.locations.edgeCacheServices.html
@@ -0,0 +1,258 @@
+
+
+
+Network Services API . projects . locations . edgeCacheServices
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+ Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being requested. See the operation documentation for the appropriate value for this field. (required)
+ options_requestedPolicyVersion: integer, Optional. The policy format version to be returned. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional bindings must specify version 3. Policies without any conditional bindings may specify any valid value or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+ Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being specified. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for `SetIamPolicy` method.
+ "policy": { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/). # REQUIRED: The complete policy to be applied to the `resource`. The size of the policy is limited to a few 10s of KB. An empty policy is a valid policy but certain Cloud Platform services (such as Projects) might reject them.
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ },
+ "updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only the fields in the mask will be modified. If no mask is provided, the following default mask is used: `paths: "bindings, etag"`
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+ Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy detail is being requested. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for `TestIamPermissions` method.
+ "permissions": [ # The set of permissions to check for the `resource`. Permissions with wildcards (such as '*' or 'storage.*') are not allowed. For more information see [IAM Overview](https://cloud.google.com/iam/docs/overview#permissions).
+ "A String",
+ ],
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response message for `TestIamPermissions` method.
+ "permissions": [ # A subset of `TestPermissionsRequest.permissions` that the caller is allowed.
+ "A String",
+ ],
+}
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/networkservices_v1.projects.locations.html b/docs/dyn/networkservices_v1.projects.locations.html
new file mode 100644
index 00000000000..669495e2e34
--- /dev/null
+++ b/docs/dyn/networkservices_v1.projects.locations.html
@@ -0,0 +1,191 @@
+
+
+
+Network Services API . projects . locations
+Instance Methods
+
+Returns the edgeCacheKeysets Resource.
+
+
+Returns the edgeCacheOrigins Resource.
+
+
+Returns the edgeCacheServices Resource.
+
+
+ operations()
+
+Returns the operations Resource.
+
+
+ close()
+Close httplib2 connections.
+
+Gets information about a location.
+
+ list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+Lists information about the supported locations for this service.
+
+ list_next(previous_request, previous_response)
+Retrieves the next page of results.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ get(name, x__xgafv=None)
+ Gets information about a location.
+
+Args:
+ name: string, Resource name for the location. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A resource that represents Google Cloud Platform location.
+ "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo".
+ "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"}
+ "a_key": "A String",
+ },
+ "locationId": "A String", # The canonical id for this location. For example: `"us-east1"`.
+ "metadata": { # Service-specific metadata. For example the available capacity at the given location.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # Resource name for the location, which may vary between implementations. For example: `"projects/example-project/locations/us-east1"`
+}
+
+
+
+ list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+ Lists information about the supported locations for this service.
+
+Args:
+ name: string, The resource that owns the locations collection, if applicable. (required)
+ filter: string, A filter to narrow down results to a preferred subset. The filtering language accepts strings like "displayName=tokyo", and is documented in more detail in [AIP-160](https://google.aip.dev/160).
+ pageSize: integer, The maximum number of results to return. If not set, the service selects a default.
+ pageToken: string, A page token received from the `next_page_token` field in the response. Send that page token to receive the subsequent page.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # The response message for Locations.ListLocations.
+ "locations": [ # A list of locations that matches the specified filter in the request.
+ { # A resource that represents Google Cloud Platform location.
+ "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo".
+ "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"}
+ "a_key": "A String",
+ },
+ "locationId": "A String", # The canonical id for this location. For example: `"us-east1"`.
+ "metadata": { # Service-specific metadata. For example the available capacity at the given location.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # Resource name for the location, which may vary between implementations. For example: `"projects/example-project/locations/us-east1"`
+ },
+ ],
+ "nextPageToken": "A String", # The standard List next-page token.
+}
+
+
+
+ list_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/networkservices_v1.projects.locations.operations.html b/docs/dyn/networkservices_v1.projects.locations.operations.html
new file mode 100644
index 00000000000..800bf6ac32e
--- /dev/null
+++ b/docs/dyn/networkservices_v1.projects.locations.operations.html
@@ -0,0 +1,235 @@
+
+
+
+Network Services API . projects . locations . operations
+Instance Methods
+
+ cancel(name, body=None, x__xgafv=None)
+Starts asynchronous cancellation on a long-running operation. The server makes a best effort to cancel the operation, but success is not guaranteed. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`. Clients can use Operations.GetOperation or other methods to check whether the cancellation succeeded or whether the operation completed despite cancellation. On successful cancellation, the operation is not deleted; instead, it becomes an operation with an Operation.error value with a google.rpc.Status.code of 1, corresponding to `Code.CANCELLED`.
+
+ close()
+Close httplib2 connections.
+
+Deletes a long-running operation. This method indicates that the client is no longer interested in the operation result. It does not cancel the operation. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`.
+
+Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+ list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `"/v1/{name=users/*}/operations"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.
+
+ list_next(previous_request, previous_response)
+Retrieves the next page of results.
+Method Details
+
+ cancel(name, body=None, x__xgafv=None)
+ Starts asynchronous cancellation on a long-running operation. The server makes a best effort to cancel the operation, but success is not guaranteed. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`. Clients can use Operations.GetOperation or other methods to check whether the cancellation succeeded or whether the operation completed despite cancellation. On successful cancellation, the operation is not deleted; instead, it becomes an operation with an Operation.error value with a google.rpc.Status.code of 1, corresponding to `Code.CANCELLED`.
+
+Args:
+ name: string, The name of the operation resource to be cancelled. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # The request message for Operations.CancelOperation.
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`.
+}
+
+
+
+ close()
+ Close httplib2 connections.
+
+
+
+ delete(name, x__xgafv=None)
+ Deletes a long-running operation. This method indicates that the client is no longer interested in the operation result. It does not cancel the operation. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`.
+
+Args:
+ name: string, The name of the operation resource to be deleted. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`.
+}
+
+
+
+ get(name, x__xgafv=None)
+ Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+ name: string, The name of the operation resource. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+
+
+
+ list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+ Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `"/v1/{name=users/*}/operations"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.
+
+Args:
+ name: string, The name of the operation's parent resource. (required)
+ filter: string, The standard list filter.
+ pageSize: integer, The standard list page size.
+ pageToken: string, The standard list page token.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # The response message for Operations.ListOperations.
+ "nextPageToken": "A String", # The standard List next-page token.
+ "operations": [ # A list of operations that matches the specified filter in the request.
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ },
+ ],
+}
+
+
+
+ list_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/networkservices_v1beta1.html b/docs/dyn/networkservices_v1beta1.html
new file mode 100644
index 00000000000..8aab358b89c
--- /dev/null
+++ b/docs/dyn/networkservices_v1beta1.html
@@ -0,0 +1,111 @@
+
+
+
+Network Services API
+Instance Methods
+
+ projects()
+
+Returns the projects Resource.
+
+
+ close()
+Close httplib2 connections.
+
+Create a BatchHttpRequest object based on the discovery document.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ new_batch_http_request()
+ Create a BatchHttpRequest object based on the discovery document.
+
+ Args:
+ callback: callable, A callback to be called for each response, of the
+ form callback(id, response, exception). The first parameter is the
+ request id, and the second is the deserialized response object. The
+ third is an apiclient.errors.HttpError exception object if an HTTP
+ error occurred while processing the request, or None if no error
+ occurred.
+
+ Returns:
+ A BatchHttpRequest object based on the discovery document.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/networkservices_v1beta1.projects.html b/docs/dyn/networkservices_v1beta1.projects.html
new file mode 100644
index 00000000000..4391ddc1c67
--- /dev/null
+++ b/docs/dyn/networkservices_v1beta1.projects.html
@@ -0,0 +1,91 @@
+
+
+
+Network Services API . projects
+Instance Methods
+
+ locations()
+
+Returns the locations Resource.
+
+
+ close()
+Close httplib2 connections.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/networkservices_v1beta1.projects.locations.endpointPolicies.html b/docs/dyn/networkservices_v1beta1.projects.locations.endpointPolicies.html
new file mode 100644
index 00000000000..f026b73d3b6
--- /dev/null
+++ b/docs/dyn/networkservices_v1beta1.projects.locations.endpointPolicies.html
@@ -0,0 +1,560 @@
+
+
+
+Network Services API . projects . locations . endpointPolicies
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+ create(parent, body=None, endpointPolicyId=None, x__xgafv=None)
+Creates a new EndpointPolicy in a given project and location.
+
+Deletes a single EndpointPolicy.
+
+Gets details of a single EndpointPolicy.
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+ list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+Lists EndpointPolicies in a given project and location.
+
+ list_next(previous_request, previous_response)
+Retrieves the next page of results.
+
+ patch(name, body=None, updateMask=None, x__xgafv=None)
+Updates the parameters of a single EndpointPolicy.
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ create(parent, body=None, endpointPolicyId=None, x__xgafv=None)
+ Creates a new EndpointPolicy in a given project and location.
+
+Args:
+ parent: string, Required. The parent resource of the EndpointPolicy. Must be in the format `projects/*/locations/global`. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # EndpointPolicy is a resource that helps apply desired configuration on the endpoints that match specific criteria. For example, this resource can be used to apply "authentication config" an all endpoints that serve on port 8080.
+ "authorizationPolicy": "A String", # Optional. This field specifies the URL of AuthorizationPolicy resource that applies authorization policies to the inbound traffic at the matched endpoints. Refer to Authorization. If this field is not specified, authorization is disabled(no authz checks) for this endpoint. Applicable only when EndpointPolicyType is SIDECAR_PROXY.
+ "clientTlsPolicy": "A String", # Optional. A URL referring to a ClientTlsPolicy resource. ClientTlsPolicy can be set to specify the authentication for traffic from the proxy to the actual endpoints. More specifically, it is applied to the outgoing traffic from the proxy to the endpoint. This is typically used for sidecar model where the proxy identifies itself as endpoint to the control plane, with the connection between sidecar and endpoint requiring authentication. If this field is not set, authentication is disabled(open). Applicable only when EndpointPolicyType is SIDECAR_PROXY.
+ "createTime": "A String", # Output only. The timestamp when the resource was created.
+ "description": "A String", # Optional. A free-text description of the resource. Max length 1024 characters.
+ "endpointMatcher": { # A definition of a matcher that selects endpoints to which the policies should be applied. # Required. A matcher that selects endpoints to which the policies should be applied.
+ "metadataLabelMatcher": { # The matcher that is based on node metadata presented by xDS clients. # The matcher is based on node metadata presented by xDS clients.
+ "metadataLabelMatchCriteria": "A String", # Specifies how matching should be done. Supported values are: MATCH_ANY: At least one of the Labels specified in the matcher should match the metadata presented by xDS client. MATCH_ALL: The metadata presented by the xDS client should contain all of the labels specified here. The selection is determined based on the best match. For example, suppose there are three EndpointPolicy resources P1, P2 and P3 and if P1 has a the matcher as MATCH_ANY , P2 has MATCH_ALL , and P3 has MATCH_ALL . If a client with label connects, the config from P1 will be selected. If a client with label connects, the config from P2 will be selected. If a client with label connects, the config from P3 will be selected. If there is more than one best match, (for example, if a config P4 with selector exists and if a client with label connects), an error will be thrown.
+ "metadataLabels": [ # The list of label value pairs that must match labels in the provided metadata based on filterMatchCriteria This list can have at most 64 entries. The list can be empty if the match criteria is MATCH_ANY, to specify a wildcard match (i.e this matches any client).
+ { # Defines a name-pair value for a single label.
+ "labelName": "A String", # Required. Label name presented as key in xDS Node Metadata.
+ "labelValue": "A String", # Required. Label value presented as value corresponding to the above key, in xDS Node Metadata.
+ },
+ ],
+ },
+ },
+ "labels": { # Optional. Set of label tags associated with the EndpointPolicy resource.
+ "a_key": "A String",
+ },
+ "name": "A String", # Required. Name of the EndpointPolicy resource. It matches pattern `projects/{project}/locations/global/endpointPolicies/{endpoint_policy}`.
+ "serverTlsPolicy": "A String", # Optional. A URL referring to ServerTlsPolicy resource. ServerTlsPolicy is used to determine the authentication policy to be applied to terminate the inbound traffic at the identified backends. If this field is not set, authentication is disabled(open) for this endpoint.
+ "trafficPortSelector": { # Specification of a port-based selector. # Optional. Port selector for the (matched) endpoints. If no port selector is provided, the matched config is applied to all ports.
+ "ports": [ # Optional. A list of ports. Can be port numbers or port range (example, [80-90] specifies all ports from 80 to 90, including 80 and 90) or named ports or * to specify all ports. If the list is empty, all ports are selected.
+ "A String",
+ ],
+ },
+ "type": "A String", # Required. The type of endpoint policy. This is primarily used to validate the configuration.
+ "updateTime": "A String", # Output only. The timestamp when the resource was updated.
+}
+
+ endpointPolicyId: string, Required. Short name of the EndpointPolicy resource to be created. E.g. "CustomECS".
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+
+
+
+ delete(name, x__xgafv=None)
+ Deletes a single EndpointPolicy.
+
+Args:
+ name: string, Required. A name of the EndpointPolicy to delete. Must be in the format `projects/*/locations/global/endpointPolicies/*`. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+
+
+
+ get(name, x__xgafv=None)
+ Gets details of a single EndpointPolicy.
+
+Args:
+ name: string, Required. A name of the EndpointPolicy to get. Must be in the format `projects/*/locations/global/endpointPolicies/*`. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # EndpointPolicy is a resource that helps apply desired configuration on the endpoints that match specific criteria. For example, this resource can be used to apply "authentication config" an all endpoints that serve on port 8080.
+ "authorizationPolicy": "A String", # Optional. This field specifies the URL of AuthorizationPolicy resource that applies authorization policies to the inbound traffic at the matched endpoints. Refer to Authorization. If this field is not specified, authorization is disabled(no authz checks) for this endpoint. Applicable only when EndpointPolicyType is SIDECAR_PROXY.
+ "clientTlsPolicy": "A String", # Optional. A URL referring to a ClientTlsPolicy resource. ClientTlsPolicy can be set to specify the authentication for traffic from the proxy to the actual endpoints. More specifically, it is applied to the outgoing traffic from the proxy to the endpoint. This is typically used for sidecar model where the proxy identifies itself as endpoint to the control plane, with the connection between sidecar and endpoint requiring authentication. If this field is not set, authentication is disabled(open). Applicable only when EndpointPolicyType is SIDECAR_PROXY.
+ "createTime": "A String", # Output only. The timestamp when the resource was created.
+ "description": "A String", # Optional. A free-text description of the resource. Max length 1024 characters.
+ "endpointMatcher": { # A definition of a matcher that selects endpoints to which the policies should be applied. # Required. A matcher that selects endpoints to which the policies should be applied.
+ "metadataLabelMatcher": { # The matcher that is based on node metadata presented by xDS clients. # The matcher is based on node metadata presented by xDS clients.
+ "metadataLabelMatchCriteria": "A String", # Specifies how matching should be done. Supported values are: MATCH_ANY: At least one of the Labels specified in the matcher should match the metadata presented by xDS client. MATCH_ALL: The metadata presented by the xDS client should contain all of the labels specified here. The selection is determined based on the best match. For example, suppose there are three EndpointPolicy resources P1, P2 and P3 and if P1 has a the matcher as MATCH_ANY , P2 has MATCH_ALL , and P3 has MATCH_ALL . If a client with label connects, the config from P1 will be selected. If a client with label connects, the config from P2 will be selected. If a client with label connects, the config from P3 will be selected. If there is more than one best match, (for example, if a config P4 with selector exists and if a client with label connects), an error will be thrown.
+ "metadataLabels": [ # The list of label value pairs that must match labels in the provided metadata based on filterMatchCriteria This list can have at most 64 entries. The list can be empty if the match criteria is MATCH_ANY, to specify a wildcard match (i.e this matches any client).
+ { # Defines a name-pair value for a single label.
+ "labelName": "A String", # Required. Label name presented as key in xDS Node Metadata.
+ "labelValue": "A String", # Required. Label value presented as value corresponding to the above key, in xDS Node Metadata.
+ },
+ ],
+ },
+ },
+ "labels": { # Optional. Set of label tags associated with the EndpointPolicy resource.
+ "a_key": "A String",
+ },
+ "name": "A String", # Required. Name of the EndpointPolicy resource. It matches pattern `projects/{project}/locations/global/endpointPolicies/{endpoint_policy}`.
+ "serverTlsPolicy": "A String", # Optional. A URL referring to ServerTlsPolicy resource. ServerTlsPolicy is used to determine the authentication policy to be applied to terminate the inbound traffic at the identified backends. If this field is not set, authentication is disabled(open) for this endpoint.
+ "trafficPortSelector": { # Specification of a port-based selector. # Optional. Port selector for the (matched) endpoints. If no port selector is provided, the matched config is applied to all ports.
+ "ports": [ # Optional. A list of ports. Can be port numbers or port range (example, [80-90] specifies all ports from 80 to 90, including 80 and 90) or named ports or * to specify all ports. If the list is empty, all ports are selected.
+ "A String",
+ ],
+ },
+ "type": "A String", # Required. The type of endpoint policy. This is primarily used to validate the configuration.
+ "updateTime": "A String", # Output only. The timestamp when the resource was updated.
+}
+
+
+
+ getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+ Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being requested. See the operation documentation for the appropriate value for this field. (required)
+ options_requestedPolicyVersion: integer, Optional. The policy format version to be returned. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional bindings must specify version 3. Policies without any conditional bindings may specify any valid value or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+ Lists EndpointPolicies in a given project and location.
+
+Args:
+ parent: string, Required. The project and location from which the EndpointPolicies should be listed, specified in the format `projects/*/locations/global`. (required)
+ pageSize: integer, Maximum number of EndpointPolicies to return per call.
+ pageToken: string, The value returned by the last `ListEndpointPoliciesResponse` Indicates that this is a continuation of a prior `ListEndpointPolicies` call, and that the system should return the next page of data.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response returned by the ListEndpointPolicies method.
+ "endpointPolicies": [ # List of EndpointPolicy resources.
+ { # EndpointPolicy is a resource that helps apply desired configuration on the endpoints that match specific criteria. For example, this resource can be used to apply "authentication config" an all endpoints that serve on port 8080.
+ "authorizationPolicy": "A String", # Optional. This field specifies the URL of AuthorizationPolicy resource that applies authorization policies to the inbound traffic at the matched endpoints. Refer to Authorization. If this field is not specified, authorization is disabled(no authz checks) for this endpoint. Applicable only when EndpointPolicyType is SIDECAR_PROXY.
+ "clientTlsPolicy": "A String", # Optional. A URL referring to a ClientTlsPolicy resource. ClientTlsPolicy can be set to specify the authentication for traffic from the proxy to the actual endpoints. More specifically, it is applied to the outgoing traffic from the proxy to the endpoint. This is typically used for sidecar model where the proxy identifies itself as endpoint to the control plane, with the connection between sidecar and endpoint requiring authentication. If this field is not set, authentication is disabled(open). Applicable only when EndpointPolicyType is SIDECAR_PROXY.
+ "createTime": "A String", # Output only. The timestamp when the resource was created.
+ "description": "A String", # Optional. A free-text description of the resource. Max length 1024 characters.
+ "endpointMatcher": { # A definition of a matcher that selects endpoints to which the policies should be applied. # Required. A matcher that selects endpoints to which the policies should be applied.
+ "metadataLabelMatcher": { # The matcher that is based on node metadata presented by xDS clients. # The matcher is based on node metadata presented by xDS clients.
+ "metadataLabelMatchCriteria": "A String", # Specifies how matching should be done. Supported values are: MATCH_ANY: At least one of the Labels specified in the matcher should match the metadata presented by xDS client. MATCH_ALL: The metadata presented by the xDS client should contain all of the labels specified here. The selection is determined based on the best match. For example, suppose there are three EndpointPolicy resources P1, P2 and P3 and if P1 has a the matcher as MATCH_ANY , P2 has MATCH_ALL , and P3 has MATCH_ALL . If a client with label connects, the config from P1 will be selected. If a client with label connects, the config from P2 will be selected. If a client with label connects, the config from P3 will be selected. If there is more than one best match, (for example, if a config P4 with selector exists and if a client with label connects), an error will be thrown.
+ "metadataLabels": [ # The list of label value pairs that must match labels in the provided metadata based on filterMatchCriteria This list can have at most 64 entries. The list can be empty if the match criteria is MATCH_ANY, to specify a wildcard match (i.e this matches any client).
+ { # Defines a name-pair value for a single label.
+ "labelName": "A String", # Required. Label name presented as key in xDS Node Metadata.
+ "labelValue": "A String", # Required. Label value presented as value corresponding to the above key, in xDS Node Metadata.
+ },
+ ],
+ },
+ },
+ "labels": { # Optional. Set of label tags associated with the EndpointPolicy resource.
+ "a_key": "A String",
+ },
+ "name": "A String", # Required. Name of the EndpointPolicy resource. It matches pattern `projects/{project}/locations/global/endpointPolicies/{endpoint_policy}`.
+ "serverTlsPolicy": "A String", # Optional. A URL referring to ServerTlsPolicy resource. ServerTlsPolicy is used to determine the authentication policy to be applied to terminate the inbound traffic at the identified backends. If this field is not set, authentication is disabled(open) for this endpoint.
+ "trafficPortSelector": { # Specification of a port-based selector. # Optional. Port selector for the (matched) endpoints. If no port selector is provided, the matched config is applied to all ports.
+ "ports": [ # Optional. A list of ports. Can be port numbers or port range (example, [80-90] specifies all ports from 80 to 90, including 80 and 90) or named ports or * to specify all ports. If the list is empty, all ports are selected.
+ "A String",
+ ],
+ },
+ "type": "A String", # Required. The type of endpoint policy. This is primarily used to validate the configuration.
+ "updateTime": "A String", # Output only. The timestamp when the resource was updated.
+ },
+ ],
+ "nextPageToken": "A String", # If there might be more results than those appearing in this response, then `next_page_token` is included. To get the next set of results, call this method again using the value of `next_page_token` as `page_token`.
+}
+
+
+
+ list_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
+ patch(name, body=None, updateMask=None, x__xgafv=None)
+ Updates the parameters of a single EndpointPolicy.
+
+Args:
+ name: string, Required. Name of the EndpointPolicy resource. It matches pattern `projects/{project}/locations/global/endpointPolicies/{endpoint_policy}`. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # EndpointPolicy is a resource that helps apply desired configuration on the endpoints that match specific criteria. For example, this resource can be used to apply "authentication config" an all endpoints that serve on port 8080.
+ "authorizationPolicy": "A String", # Optional. This field specifies the URL of AuthorizationPolicy resource that applies authorization policies to the inbound traffic at the matched endpoints. Refer to Authorization. If this field is not specified, authorization is disabled(no authz checks) for this endpoint. Applicable only when EndpointPolicyType is SIDECAR_PROXY.
+ "clientTlsPolicy": "A String", # Optional. A URL referring to a ClientTlsPolicy resource. ClientTlsPolicy can be set to specify the authentication for traffic from the proxy to the actual endpoints. More specifically, it is applied to the outgoing traffic from the proxy to the endpoint. This is typically used for sidecar model where the proxy identifies itself as endpoint to the control plane, with the connection between sidecar and endpoint requiring authentication. If this field is not set, authentication is disabled(open). Applicable only when EndpointPolicyType is SIDECAR_PROXY.
+ "createTime": "A String", # Output only. The timestamp when the resource was created.
+ "description": "A String", # Optional. A free-text description of the resource. Max length 1024 characters.
+ "endpointMatcher": { # A definition of a matcher that selects endpoints to which the policies should be applied. # Required. A matcher that selects endpoints to which the policies should be applied.
+ "metadataLabelMatcher": { # The matcher that is based on node metadata presented by xDS clients. # The matcher is based on node metadata presented by xDS clients.
+ "metadataLabelMatchCriteria": "A String", # Specifies how matching should be done. Supported values are: MATCH_ANY: At least one of the Labels specified in the matcher should match the metadata presented by xDS client. MATCH_ALL: The metadata presented by the xDS client should contain all of the labels specified here. The selection is determined based on the best match. For example, suppose there are three EndpointPolicy resources P1, P2 and P3 and if P1 has a the matcher as MATCH_ANY , P2 has MATCH_ALL , and P3 has MATCH_ALL . If a client with label connects, the config from P1 will be selected. If a client with label connects, the config from P2 will be selected. If a client with label connects, the config from P3 will be selected. If there is more than one best match, (for example, if a config P4 with selector exists and if a client with label connects), an error will be thrown.
+ "metadataLabels": [ # The list of label value pairs that must match labels in the provided metadata based on filterMatchCriteria This list can have at most 64 entries. The list can be empty if the match criteria is MATCH_ANY, to specify a wildcard match (i.e this matches any client).
+ { # Defines a name-pair value for a single label.
+ "labelName": "A String", # Required. Label name presented as key in xDS Node Metadata.
+ "labelValue": "A String", # Required. Label value presented as value corresponding to the above key, in xDS Node Metadata.
+ },
+ ],
+ },
+ },
+ "labels": { # Optional. Set of label tags associated with the EndpointPolicy resource.
+ "a_key": "A String",
+ },
+ "name": "A String", # Required. Name of the EndpointPolicy resource. It matches pattern `projects/{project}/locations/global/endpointPolicies/{endpoint_policy}`.
+ "serverTlsPolicy": "A String", # Optional. A URL referring to ServerTlsPolicy resource. ServerTlsPolicy is used to determine the authentication policy to be applied to terminate the inbound traffic at the identified backends. If this field is not set, authentication is disabled(open) for this endpoint.
+ "trafficPortSelector": { # Specification of a port-based selector. # Optional. Port selector for the (matched) endpoints. If no port selector is provided, the matched config is applied to all ports.
+ "ports": [ # Optional. A list of ports. Can be port numbers or port range (example, [80-90] specifies all ports from 80 to 90, including 80 and 90) or named ports or * to specify all ports. If the list is empty, all ports are selected.
+ "A String",
+ ],
+ },
+ "type": "A String", # Required. The type of endpoint policy. This is primarily used to validate the configuration.
+ "updateTime": "A String", # Output only. The timestamp when the resource was updated.
+}
+
+ updateMask: string, Optional. Field mask is used to specify the fields to be overwritten in the EndpointPolicy resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then all fields will be overwritten.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+
+
+
+ setIamPolicy(resource, body=None, x__xgafv=None)
+ Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy is being specified. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for `SetIamPolicy` method.
+ "policy": { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/). # REQUIRED: The complete policy to be applied to the `resource`. The size of the policy is limited to a few 10s of KB. An empty policy is a valid policy but certain Cloud Platform services (such as Projects) might reject them.
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ },
+ "updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only the fields in the mask will be modified. If no mask is provided, the following default mask is used: `paths: "bindings, etag"`
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members` to a single `role`. Members can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/).
+ "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+ { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging.
+ "auditLogConfigs": [ # The configuration for logging of each type of permission.
+ { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+ "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+ "A String",
+ ],
+ "logType": "A String", # The log type that this config enables.
+ },
+ ],
+ "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+ },
+ ],
+ "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one member.
+ { # Associates `members` with a `role`.
+ "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the members in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+ "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+ "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+ "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+ "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+ },
+ "members": [ # Specifies the identities requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`.
+ "A String",
+ ],
+ "role": "A String", # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
+ },
+ ],
+ "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+ "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
+
+ testIamPermissions(resource, body=None, x__xgafv=None)
+ Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+
+Args:
+ resource: string, REQUIRED: The resource for which the policy detail is being requested. See the operation documentation for the appropriate value for this field. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for `TestIamPermissions` method.
+ "permissions": [ # The set of permissions to check for the `resource`. Permissions with wildcards (such as '*' or 'storage.*') are not allowed. For more information see [IAM Overview](https://cloud.google.com/iam/docs/overview#permissions).
+ "A String",
+ ],
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response message for `TestIamPermissions` method.
+ "permissions": [ # A subset of `TestPermissionsRequest.permissions` that the caller is allowed.
+ "A String",
+ ],
+}
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/networkservices_v1beta1.projects.locations.html b/docs/dyn/networkservices_v1beta1.projects.locations.html
new file mode 100644
index 00000000000..93086a64a71
--- /dev/null
+++ b/docs/dyn/networkservices_v1beta1.projects.locations.html
@@ -0,0 +1,181 @@
+
+
+
+Network Services API . projects . locations
+Instance Methods
+
+Returns the endpointPolicies Resource.
+
+
+ operations()
+
+Returns the operations Resource.
+
+
+ close()
+Close httplib2 connections.
+
+Gets information about a location.
+
+ list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+Lists information about the supported locations for this service.
+
+ list_next(previous_request, previous_response)
+Retrieves the next page of results.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ get(name, x__xgafv=None)
+ Gets information about a location.
+
+Args:
+ name: string, Resource name for the location. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A resource that represents Google Cloud Platform location.
+ "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo".
+ "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"}
+ "a_key": "A String",
+ },
+ "locationId": "A String", # The canonical id for this location. For example: `"us-east1"`.
+ "metadata": { # Service-specific metadata. For example the available capacity at the given location.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # Resource name for the location, which may vary between implementations. For example: `"projects/example-project/locations/us-east1"`
+}
+
+
+
+ list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+ Lists information about the supported locations for this service.
+
+Args:
+ name: string, The resource that owns the locations collection, if applicable. (required)
+ filter: string, A filter to narrow down results to a preferred subset. The filtering language accepts strings like "displayName=tokyo", and is documented in more detail in [AIP-160](https://google.aip.dev/160).
+ pageSize: integer, The maximum number of results to return. If not set, the service selects a default.
+ pageToken: string, A page token received from the `next_page_token` field in the response. Send that page token to receive the subsequent page.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # The response message for Locations.ListLocations.
+ "locations": [ # A list of locations that matches the specified filter in the request.
+ { # A resource that represents Google Cloud Platform location.
+ "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo".
+ "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"}
+ "a_key": "A String",
+ },
+ "locationId": "A String", # The canonical id for this location. For example: `"us-east1"`.
+ "metadata": { # Service-specific metadata. For example the available capacity at the given location.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # Resource name for the location, which may vary between implementations. For example: `"projects/example-project/locations/us-east1"`
+ },
+ ],
+ "nextPageToken": "A String", # The standard List next-page token.
+}
+
+
+
+ list_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/networkservices_v1beta1.projects.locations.operations.html b/docs/dyn/networkservices_v1beta1.projects.locations.operations.html
new file mode 100644
index 00000000000..a9c4de8a5b7
--- /dev/null
+++ b/docs/dyn/networkservices_v1beta1.projects.locations.operations.html
@@ -0,0 +1,235 @@
+
+
+
+Network Services API . projects . locations . operations
+Instance Methods
+
+ cancel(name, body=None, x__xgafv=None)
+Starts asynchronous cancellation on a long-running operation. The server makes a best effort to cancel the operation, but success is not guaranteed. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`. Clients can use Operations.GetOperation or other methods to check whether the cancellation succeeded or whether the operation completed despite cancellation. On successful cancellation, the operation is not deleted; instead, it becomes an operation with an Operation.error value with a google.rpc.Status.code of 1, corresponding to `Code.CANCELLED`.
+
+ close()
+Close httplib2 connections.
+
+Deletes a long-running operation. This method indicates that the client is no longer interested in the operation result. It does not cancel the operation. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`.
+
+Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+ list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `"/v1/{name=users/*}/operations"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.
+
+ list_next(previous_request, previous_response)
+Retrieves the next page of results.
+Method Details
+
+ cancel(name, body=None, x__xgafv=None)
+ Starts asynchronous cancellation on a long-running operation. The server makes a best effort to cancel the operation, but success is not guaranteed. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`. Clients can use Operations.GetOperation or other methods to check whether the cancellation succeeded or whether the operation completed despite cancellation. On successful cancellation, the operation is not deleted; instead, it becomes an operation with an Operation.error value with a google.rpc.Status.code of 1, corresponding to `Code.CANCELLED`.
+
+Args:
+ name: string, The name of the operation resource to be cancelled. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # The request message for Operations.CancelOperation.
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`.
+}
+
+
+
+ close()
+ Close httplib2 connections.
+
+
+
+ delete(name, x__xgafv=None)
+ Deletes a long-running operation. This method indicates that the client is no longer interested in the operation result. It does not cancel the operation. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`.
+
+Args:
+ name: string, The name of the operation resource to be deleted. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`.
+}
+
+
+
+ get(name, x__xgafv=None)
+ Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+ name: string, The name of the operation resource. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+
+
+
+ list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+ Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `"/v1/{name=users/*}/operations"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.
+
+Args:
+ name: string, The name of the operation's parent resource. (required)
+ filter: string, The standard list filter.
+ pageSize: integer, The standard list page size.
+ pageToken: string, The standard list page token.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # The response message for Operations.ListOperations.
+ "nextPageToken": "A String", # The standard List next-page token.
+ "operations": [ # A list of operations that matches the specified filter in the request.
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ },
+ ],
+}
+
+
+
+ list_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/notebooks_v1.projects.locations.executions.html b/docs/dyn/notebooks_v1.projects.locations.executions.html
index a64bf7930c8..5df743e47d1 100644
--- a/docs/dyn/notebooks_v1.projects.locations.executions.html
+++ b/docs/dyn/notebooks_v1.projects.locations.executions.html
@@ -205,7 +205,7 @@ Method Details
Gets details of executions
Args:
- name: string, Required. Format: `projects/{project_id}/locations/{location}/schedules/{execution_id}` (required)
+ name: string, Required. Format: `projects/{project_id}/locations/{location}/executions/{execution_id}` (required)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
diff --git a/docs/dyn/ondemandscanning_v1.projects.locations.scans.vulnerabilities.html b/docs/dyn/ondemandscanning_v1.projects.locations.scans.vulnerabilities.html
index 9dfe9f116c1..023892c5ebf 100644
--- a/docs/dyn/ondemandscanning_v1.projects.locations.scans.vulnerabilities.html
+++ b/docs/dyn/ondemandscanning_v1.projects.locations.scans.vulnerabilities.html
@@ -124,6 +124,36 @@ Method Details
],
},
"build": { # Details of a build occurrence. # Describes a verifiable build.
+ "intotoProvenance": { # In-toto Provenance representation as defined in spec.
+ "builderConfig": { # required
+ "id": "A String",
+ },
+ "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty.
+ "A String",
+ ],
+ "metadata": { # Other properties of the build.
+ "buildFinishedOn": "A String", # The timestamp of when the build completed.
+ "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec.
+ "buildStartedOn": "A String", # The timestamp of when the build started.
+ "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete.
+ "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe.
+ "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete.
+ "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic".
+ },
+ "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output.
+ },
+ "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required
+ "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint.
+ "A String",
+ ],
+ "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64.
+ "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use.
+ "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy.
+ "a_key": "A String",
+ },
+ "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials.
+ },
+ },
"provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # Required. The actual provenance for the build.
"buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details.
"a_key": "A String",
@@ -281,6 +311,70 @@ Method Details
"cpe": "A String", # The CPE of the resource being scanned.
"lastScanTime": "A String", # The last time this resource was scanned.
},
+ "dsseAttestation": { # Describes an attestation of an artifact using dsse.
+ "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # If doing something security critical, make sure to verify the signatures in this metadata.
+ "payload": "A String",
+ "payloadType": "A String",
+ "signatures": [
+ {
+ "keyid": "A String",
+ "sig": "A String",
+ },
+ ],
+ },
+ "statement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json".
+ "predicateType": "A String", # "https://in-toto.io/Provenance/v0.1" for InTotoProvenance.
+ "provenance": {
+ "builderConfig": { # required
+ "id": "A String",
+ },
+ "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty.
+ "A String",
+ ],
+ "metadata": { # Other properties of the build.
+ "buildFinishedOn": "A String", # The timestamp of when the build completed.
+ "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec.
+ "buildStartedOn": "A String", # The timestamp of when the build started.
+ "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete.
+ "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe.
+ "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete.
+ "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic".
+ },
+ "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output.
+ },
+ "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required
+ "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint.
+ "A String",
+ ],
+ "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64.
+ "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use.
+ "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy.
+ "a_key": "A String",
+ },
+ "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials.
+ },
+ },
+ "subject": [
+ {
+ "digest": { # "": ""
+ "a_key": "A String",
+ },
+ "name": "A String",
+ },
+ ],
+ "type": "A String", # Always "https://in-toto.io/Statement/v0.1".
+ },
+ },
+ "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # https://github.com/secure-systems-lab/dsse
+ "payload": "A String",
+ "payloadType": "A String",
+ "signatures": [
+ {
+ "keyid": "A String",
+ "sig": "A String",
+ },
+ ],
+ },
"image": { # Details of the derived image portion of the DockerImage relationship. This image would be produced from a Dockerfile with FROM . # Describes how this resource derives from the basis in the associated note.
"baseResourceUrl": "A String", # Output only. This contains the base image URL for the derived image occurrence.
"distance": 42, # Output only. The number of layers by which this image differs from the associated image basis.
diff --git a/docs/dyn/ondemandscanning_v1beta1.projects.locations.scans.vulnerabilities.html b/docs/dyn/ondemandscanning_v1beta1.projects.locations.scans.vulnerabilities.html
index e1430485e88..d9f89136a61 100644
--- a/docs/dyn/ondemandscanning_v1beta1.projects.locations.scans.vulnerabilities.html
+++ b/docs/dyn/ondemandscanning_v1beta1.projects.locations.scans.vulnerabilities.html
@@ -124,6 +124,36 @@ Method Details
],
},
"build": { # Details of a build occurrence. # Describes a verifiable build.
+ "intotoProvenance": { # In-toto Provenance representation as defined in spec.
+ "builderConfig": { # required
+ "id": "A String",
+ },
+ "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty.
+ "A String",
+ ],
+ "metadata": { # Other properties of the build.
+ "buildFinishedOn": "A String", # The timestamp of when the build completed.
+ "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec.
+ "buildStartedOn": "A String", # The timestamp of when the build started.
+ "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete.
+ "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe.
+ "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete.
+ "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic".
+ },
+ "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output.
+ },
+ "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required
+ "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint.
+ "A String",
+ ],
+ "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64.
+ "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use.
+ "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy.
+ "a_key": "A String",
+ },
+ "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials.
+ },
+ },
"provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # Required. The actual provenance for the build.
"buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details.
"a_key": "A String",
@@ -281,6 +311,70 @@ Method Details
"cpe": "A String", # The CPE of the resource being scanned.
"lastScanTime": "A String", # The last time this resource was scanned.
},
+ "dsseAttestation": { # Describes an attestation of an artifact using dsse.
+ "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # If doing something security critical, make sure to verify the signatures in this metadata.
+ "payload": "A String",
+ "payloadType": "A String",
+ "signatures": [
+ {
+ "keyid": "A String",
+ "sig": "A String",
+ },
+ ],
+ },
+ "statement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json".
+ "predicateType": "A String", # "https://in-toto.io/Provenance/v0.1" for InTotoProvenance.
+ "provenance": {
+ "builderConfig": { # required
+ "id": "A String",
+ },
+ "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty.
+ "A String",
+ ],
+ "metadata": { # Other properties of the build.
+ "buildFinishedOn": "A String", # The timestamp of when the build completed.
+ "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec.
+ "buildStartedOn": "A String", # The timestamp of when the build started.
+ "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete.
+ "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe.
+ "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete.
+ "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic".
+ },
+ "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output.
+ },
+ "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required
+ "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint.
+ "A String",
+ ],
+ "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64.
+ "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use.
+ "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy.
+ "a_key": "A String",
+ },
+ "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials.
+ },
+ },
+ "subject": [
+ {
+ "digest": { # "": ""
+ "a_key": "A String",
+ },
+ "name": "A String",
+ },
+ ],
+ "type": "A String", # Always "https://in-toto.io/Statement/v0.1".
+ },
+ },
+ "envelope": { # MUST match https://github.com/secure-systems-lab/dsse/blob/master/envelope.proto. An authenticated message of arbitrary type. # https://github.com/secure-systems-lab/dsse
+ "payload": "A String",
+ "payloadType": "A String",
+ "signatures": [
+ {
+ "keyid": "A String",
+ "sig": "A String",
+ },
+ ],
+ },
"image": { # Details of the derived image portion of the DockerImage relationship. This image would be produced from a Dockerfile with FROM . # Describes how this resource derives from the basis in the associated note.
"baseResourceUrl": "A String", # Output only. This contains the base image URL for the derived image occurrence.
"distance": 42, # Output only. The number of layers by which this image differs from the associated image basis.
diff --git a/docs/dyn/osconfig_v1.projects.html b/docs/dyn/osconfig_v1.projects.html
index 693c11934be..92de89797d5 100644
--- a/docs/dyn/osconfig_v1.projects.html
+++ b/docs/dyn/osconfig_v1.projects.html
@@ -74,6 +74,11 @@
OS Config API . projects
Instance Methods
+
+ locations()
+
+Returns the locations Resource.
+
diff --git a/docs/dyn/osconfig_v1.projects.locations.html b/docs/dyn/osconfig_v1.projects.locations.html
new file mode 100644
index 00000000000..94c8b165316
--- /dev/null
+++ b/docs/dyn/osconfig_v1.projects.locations.html
@@ -0,0 +1,91 @@
+
+
+
+OS Config API . projects . locations
+Instance Methods
+
+ instances()
+
+Returns the instances Resource.
+
+
+ close()
+Close httplib2 connections.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/osconfig_v1.projects.locations.instances.html b/docs/dyn/osconfig_v1.projects.locations.instances.html
new file mode 100644
index 00000000000..ec1b1035c83
--- /dev/null
+++ b/docs/dyn/osconfig_v1.projects.locations.instances.html
@@ -0,0 +1,96 @@
+
+
+
+OS Config API . projects . locations . instances
+Instance Methods
+
+ inventories()
+
+Returns the inventories Resource.
+
+
+Returns the vulnerabilityReports Resource.
+
+
+ close()
+Close httplib2 connections.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/osconfig_v1.projects.locations.instances.inventories.html b/docs/dyn/osconfig_v1.projects.locations.instances.inventories.html
new file mode 100644
index 00000000000..4c4bf04324d
--- /dev/null
+++ b/docs/dyn/osconfig_v1.projects.locations.instances.inventories.html
@@ -0,0 +1,441 @@
+
+
+
+OS Config API . projects . locations . instances . inventories
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+ get(name, view=None, x__xgafv=None)
+Get inventory data for the specified VM instance. If the VM has no associated inventory, the message `NOT_FOUND` is returned.
+
+ list(parent, filter=None, pageSize=None, pageToken=None, view=None, x__xgafv=None)
+List inventory data for all VM instances in the specified zone.
+
+ list_next(previous_request, previous_response)
+Retrieves the next page of results.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ get(name, view=None, x__xgafv=None)
+ Get inventory data for the specified VM instance. If the VM has no associated inventory, the message `NOT_FOUND` is returned.
+
+Args:
+ name: string, Required. API resource name for inventory resource. Format: `projects/{project}/locations/{location}/instances/{instance}/inventory` For `{project}`, either `project-number` or `project-id` can be provided. For `{instance}`, either Compute Engine `instance-id` or `instance-name` can be provided. (required)
+ view: string, Inventory view indicating what information should be included in the inventory resource. If unspecified, the default view is BASIC.
+ Allowed values
+ INVENTORY_VIEW_UNSPECIFIED - The default value. The API defaults to the BASIC view.
+ BASIC - Returns the basic inventory information that includes `os_info`.
+ FULL - Returns all fields.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This API resource represents the available inventory data for a Compute Engine virtual machine (VM) instance at a given point in time. You can use this API resource to determine the inventory data of your VM. For more information, see [Information provided by OS inventory management](https://cloud.google.com/compute/docs/instances/os-inventory-management#data-collected).
+ "items": { # Inventory items related to the VM keyed by an opaque unique identifier for each inventory item. The identifier is unique to each distinct and addressable inventory item and will change, when there is a new package version.
+ "a_key": { # A single piece of inventory on a VM.
+ "availablePackage": { # Software package information of the operating system. # Software package available to be installed on the VM instance.
+ "aptPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of an APT package. For details about the apt package manager, see https://wiki.debian.org/Apt.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "cosPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a COS package.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "googetPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a Googet package. For details about the googet package manager, see https://github.com/google/googet.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "qfePackage": { # Information related to a Quick Fix Engineering package. Fields are taken from Windows QuickFixEngineering Interface and match the source names: https://docs.microsoft.com/en-us/windows/win32/cimwin32prov/win32-quickfixengineering # Details of a Windows Quick Fix engineering package. See https://docs.microsoft.com/en-us/windows/win32/cimwin32prov/win32-quickfixengineering for info in Windows Quick Fix Engineering.
+ "caption": "A String", # A short textual description of the QFE update.
+ "description": "A String", # A textual description of the QFE update.
+ "hotFixId": "A String", # Unique identifier associated with a particular QFE update.
+ "installTime": "A String", # Date that the QFE update was installed. Mapped from installed_on field.
+ },
+ "wuaPackage": { # Details related to a Windows Update package. Field data and names are taken from Windows Update API IUpdate Interface: https://docs.microsoft.com/en-us/windows/win32/api/_wua/ Descriptive fields like title, and description are localized based on the locale of the VM being updated. # Details of a Windows Update package. See https://docs.microsoft.com/en-us/windows/win32/api/_wua/ for information about Windows Update.
+ "categories": [ # The categories that are associated with this update package.
+ { # Categories specified by the Windows Update.
+ "id": "A String", # The identifier of the windows update category.
+ "name": "A String", # The name of the windows update category.
+ },
+ ],
+ "description": "A String", # The localized description of the update package.
+ "kbArticleIds": [ # A collection of Microsoft Knowledge Base article IDs that are associated with the update package.
+ "A String",
+ ],
+ "lastDeploymentChangeTime": "A String", # The last published date of the update, in (UTC) date and time.
+ "moreInfoUrls": [ # A collection of URLs that provide more information about the update package.
+ "A String",
+ ],
+ "revisionNumber": 42, # The revision number of this update package.
+ "supportUrl": "A String", # A hyperlink to the language-specific support information for the update.
+ "title": "A String", # The localized title of the update package.
+ "updateId": "A String", # Gets the identifier of an update package. Stays the same across revisions.
+ },
+ "yumPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Yum package info. For details about the yum package manager, see https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/6/html/deployment_guide/ch-yum.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "zypperPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a Zypper package. For details about the Zypper package manager, see https://en.opensuse.org/SDB:Zypper_manual.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "zypperPatch": { # Details related to a Zypper Patch. # Details of a Zypper patch. For details about the Zypper package manager, see https://en.opensuse.org/SDB:Zypper_manual.
+ "category": "A String", # The category of the patch.
+ "patchName": "A String", # The name of the patch.
+ "severity": "A String", # The severity specified for this patch
+ "summary": "A String", # Any summary information provided about this patch.
+ },
+ },
+ "createTime": "A String", # When this inventory item was first detected.
+ "id": "A String", # Identifier for this item, unique across items for this VM.
+ "installedPackage": { # Software package information of the operating system. # Software package present on the VM instance.
+ "aptPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of an APT package. For details about the apt package manager, see https://wiki.debian.org/Apt.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "cosPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a COS package.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "googetPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a Googet package. For details about the googet package manager, see https://github.com/google/googet.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "qfePackage": { # Information related to a Quick Fix Engineering package. Fields are taken from Windows QuickFixEngineering Interface and match the source names: https://docs.microsoft.com/en-us/windows/win32/cimwin32prov/win32-quickfixengineering # Details of a Windows Quick Fix engineering package. See https://docs.microsoft.com/en-us/windows/win32/cimwin32prov/win32-quickfixengineering for info in Windows Quick Fix Engineering.
+ "caption": "A String", # A short textual description of the QFE update.
+ "description": "A String", # A textual description of the QFE update.
+ "hotFixId": "A String", # Unique identifier associated with a particular QFE update.
+ "installTime": "A String", # Date that the QFE update was installed. Mapped from installed_on field.
+ },
+ "wuaPackage": { # Details related to a Windows Update package. Field data and names are taken from Windows Update API IUpdate Interface: https://docs.microsoft.com/en-us/windows/win32/api/_wua/ Descriptive fields like title, and description are localized based on the locale of the VM being updated. # Details of a Windows Update package. See https://docs.microsoft.com/en-us/windows/win32/api/_wua/ for information about Windows Update.
+ "categories": [ # The categories that are associated with this update package.
+ { # Categories specified by the Windows Update.
+ "id": "A String", # The identifier of the windows update category.
+ "name": "A String", # The name of the windows update category.
+ },
+ ],
+ "description": "A String", # The localized description of the update package.
+ "kbArticleIds": [ # A collection of Microsoft Knowledge Base article IDs that are associated with the update package.
+ "A String",
+ ],
+ "lastDeploymentChangeTime": "A String", # The last published date of the update, in (UTC) date and time.
+ "moreInfoUrls": [ # A collection of URLs that provide more information about the update package.
+ "A String",
+ ],
+ "revisionNumber": 42, # The revision number of this update package.
+ "supportUrl": "A String", # A hyperlink to the language-specific support information for the update.
+ "title": "A String", # The localized title of the update package.
+ "updateId": "A String", # Gets the identifier of an update package. Stays the same across revisions.
+ },
+ "yumPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Yum package info. For details about the yum package manager, see https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/6/html/deployment_guide/ch-yum.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "zypperPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a Zypper package. For details about the Zypper package manager, see https://en.opensuse.org/SDB:Zypper_manual.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "zypperPatch": { # Details related to a Zypper Patch. # Details of a Zypper patch. For details about the Zypper package manager, see https://en.opensuse.org/SDB:Zypper_manual.
+ "category": "A String", # The category of the patch.
+ "patchName": "A String", # The name of the patch.
+ "severity": "A String", # The severity specified for this patch
+ "summary": "A String", # Any summary information provided about this patch.
+ },
+ },
+ "originType": "A String", # The origin of this inventory item.
+ "type": "A String", # The specific type of inventory, correlating to its specific details.
+ "updateTime": "A String", # When this inventory item was last modified.
+ },
+ },
+ "name": "A String", # Output only. The `Inventory` API resource name. Format: `projects/{project_number}/locations/{location}/instances/{instance_id}/inventory`
+ "osInfo": { # Operating system information for the VM. # Base level operating system information for the VM.
+ "architecture": "A String", # The system architecture of the operating system.
+ "hostname": "A String", # The VM hostname.
+ "kernelRelease": "A String", # The kernel release of the operating system.
+ "kernelVersion": "A String", # The kernel version of the operating system.
+ "longName": "A String", # The operating system long name. For example 'Debian GNU/Linux 9' or 'Microsoft Window Server 2019 Datacenter'.
+ "osconfigAgentVersion": "A String", # The current version of the OS Config agent running on the VM.
+ "shortName": "A String", # The operating system short name. For example, 'windows' or 'debian'.
+ "version": "A String", # The version of the operating system.
+ },
+ "updateTime": "A String", # Output only. Timestamp of the last reported inventory for the VM.
+}
+
+
+
+ list(parent, filter=None, pageSize=None, pageToken=None, view=None, x__xgafv=None)
+ List inventory data for all VM instances in the specified zone.
+
+Args:
+ parent: string, Required. The parent resource name. Format: `projects/{project}/locations/{location}/instances/-` For `{project}`, either `project-number` or `project-id` can be provided. (required)
+ filter: string, If provided, this field specifies the criteria that must be met by a `Inventory` API resource to be included in the response.
+ pageSize: integer, The maximum number of results to return.
+ pageToken: string, A pagination token returned from a previous call to `ListInventories` that indicates where this listing should continue from.
+ view: string, Inventory view indicating what information should be included in the inventory resource. If unspecified, the default view is BASIC.
+ Allowed values
+ INVENTORY_VIEW_UNSPECIFIED - The default value. The API defaults to the BASIC view.
+ BASIC - Returns the basic inventory information that includes `os_info`.
+ FULL - Returns all fields.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A response message for listing inventory data for all VMs in a specified location.
+ "inventories": [ # List of inventory objects.
+ { # This API resource represents the available inventory data for a Compute Engine virtual machine (VM) instance at a given point in time. You can use this API resource to determine the inventory data of your VM. For more information, see [Information provided by OS inventory management](https://cloud.google.com/compute/docs/instances/os-inventory-management#data-collected).
+ "items": { # Inventory items related to the VM keyed by an opaque unique identifier for each inventory item. The identifier is unique to each distinct and addressable inventory item and will change, when there is a new package version.
+ "a_key": { # A single piece of inventory on a VM.
+ "availablePackage": { # Software package information of the operating system. # Software package available to be installed on the VM instance.
+ "aptPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of an APT package. For details about the apt package manager, see https://wiki.debian.org/Apt.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "cosPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a COS package.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "googetPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a Googet package. For details about the googet package manager, see https://github.com/google/googet.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "qfePackage": { # Information related to a Quick Fix Engineering package. Fields are taken from Windows QuickFixEngineering Interface and match the source names: https://docs.microsoft.com/en-us/windows/win32/cimwin32prov/win32-quickfixengineering # Details of a Windows Quick Fix engineering package. See https://docs.microsoft.com/en-us/windows/win32/cimwin32prov/win32-quickfixengineering for info in Windows Quick Fix Engineering.
+ "caption": "A String", # A short textual description of the QFE update.
+ "description": "A String", # A textual description of the QFE update.
+ "hotFixId": "A String", # Unique identifier associated with a particular QFE update.
+ "installTime": "A String", # Date that the QFE update was installed. Mapped from installed_on field.
+ },
+ "wuaPackage": { # Details related to a Windows Update package. Field data and names are taken from Windows Update API IUpdate Interface: https://docs.microsoft.com/en-us/windows/win32/api/_wua/ Descriptive fields like title, and description are localized based on the locale of the VM being updated. # Details of a Windows Update package. See https://docs.microsoft.com/en-us/windows/win32/api/_wua/ for information about Windows Update.
+ "categories": [ # The categories that are associated with this update package.
+ { # Categories specified by the Windows Update.
+ "id": "A String", # The identifier of the windows update category.
+ "name": "A String", # The name of the windows update category.
+ },
+ ],
+ "description": "A String", # The localized description of the update package.
+ "kbArticleIds": [ # A collection of Microsoft Knowledge Base article IDs that are associated with the update package.
+ "A String",
+ ],
+ "lastDeploymentChangeTime": "A String", # The last published date of the update, in (UTC) date and time.
+ "moreInfoUrls": [ # A collection of URLs that provide more information about the update package.
+ "A String",
+ ],
+ "revisionNumber": 42, # The revision number of this update package.
+ "supportUrl": "A String", # A hyperlink to the language-specific support information for the update.
+ "title": "A String", # The localized title of the update package.
+ "updateId": "A String", # Gets the identifier of an update package. Stays the same across revisions.
+ },
+ "yumPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Yum package info. For details about the yum package manager, see https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/6/html/deployment_guide/ch-yum.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "zypperPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a Zypper package. For details about the Zypper package manager, see https://en.opensuse.org/SDB:Zypper_manual.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "zypperPatch": { # Details related to a Zypper Patch. # Details of a Zypper patch. For details about the Zypper package manager, see https://en.opensuse.org/SDB:Zypper_manual.
+ "category": "A String", # The category of the patch.
+ "patchName": "A String", # The name of the patch.
+ "severity": "A String", # The severity specified for this patch
+ "summary": "A String", # Any summary information provided about this patch.
+ },
+ },
+ "createTime": "A String", # When this inventory item was first detected.
+ "id": "A String", # Identifier for this item, unique across items for this VM.
+ "installedPackage": { # Software package information of the operating system. # Software package present on the VM instance.
+ "aptPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of an APT package. For details about the apt package manager, see https://wiki.debian.org/Apt.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "cosPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a COS package.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "googetPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a Googet package. For details about the googet package manager, see https://github.com/google/googet.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "qfePackage": { # Information related to a Quick Fix Engineering package. Fields are taken from Windows QuickFixEngineering Interface and match the source names: https://docs.microsoft.com/en-us/windows/win32/cimwin32prov/win32-quickfixengineering # Details of a Windows Quick Fix engineering package. See https://docs.microsoft.com/en-us/windows/win32/cimwin32prov/win32-quickfixengineering for info in Windows Quick Fix Engineering.
+ "caption": "A String", # A short textual description of the QFE update.
+ "description": "A String", # A textual description of the QFE update.
+ "hotFixId": "A String", # Unique identifier associated with a particular QFE update.
+ "installTime": "A String", # Date that the QFE update was installed. Mapped from installed_on field.
+ },
+ "wuaPackage": { # Details related to a Windows Update package. Field data and names are taken from Windows Update API IUpdate Interface: https://docs.microsoft.com/en-us/windows/win32/api/_wua/ Descriptive fields like title, and description are localized based on the locale of the VM being updated. # Details of a Windows Update package. See https://docs.microsoft.com/en-us/windows/win32/api/_wua/ for information about Windows Update.
+ "categories": [ # The categories that are associated with this update package.
+ { # Categories specified by the Windows Update.
+ "id": "A String", # The identifier of the windows update category.
+ "name": "A String", # The name of the windows update category.
+ },
+ ],
+ "description": "A String", # The localized description of the update package.
+ "kbArticleIds": [ # A collection of Microsoft Knowledge Base article IDs that are associated with the update package.
+ "A String",
+ ],
+ "lastDeploymentChangeTime": "A String", # The last published date of the update, in (UTC) date and time.
+ "moreInfoUrls": [ # A collection of URLs that provide more information about the update package.
+ "A String",
+ ],
+ "revisionNumber": 42, # The revision number of this update package.
+ "supportUrl": "A String", # A hyperlink to the language-specific support information for the update.
+ "title": "A String", # The localized title of the update package.
+ "updateId": "A String", # Gets the identifier of an update package. Stays the same across revisions.
+ },
+ "yumPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Yum package info. For details about the yum package manager, see https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/6/html/deployment_guide/ch-yum.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "zypperPackage": { # Information related to the a standard versioned package. This includes package info for APT, Yum, Zypper, and Googet package managers. # Details of a Zypper package. For details about the Zypper package manager, see https://en.opensuse.org/SDB:Zypper_manual.
+ "architecture": "A String", # The system architecture this package is intended for.
+ "packageName": "A String", # The name of the package.
+ "version": "A String", # The version of the package.
+ },
+ "zypperPatch": { # Details related to a Zypper Patch. # Details of a Zypper patch. For details about the Zypper package manager, see https://en.opensuse.org/SDB:Zypper_manual.
+ "category": "A String", # The category of the patch.
+ "patchName": "A String", # The name of the patch.
+ "severity": "A String", # The severity specified for this patch
+ "summary": "A String", # Any summary information provided about this patch.
+ },
+ },
+ "originType": "A String", # The origin of this inventory item.
+ "type": "A String", # The specific type of inventory, correlating to its specific details.
+ "updateTime": "A String", # When this inventory item was last modified.
+ },
+ },
+ "name": "A String", # Output only. The `Inventory` API resource name. Format: `projects/{project_number}/locations/{location}/instances/{instance_id}/inventory`
+ "osInfo": { # Operating system information for the VM. # Base level operating system information for the VM.
+ "architecture": "A String", # The system architecture of the operating system.
+ "hostname": "A String", # The VM hostname.
+ "kernelRelease": "A String", # The kernel release of the operating system.
+ "kernelVersion": "A String", # The kernel version of the operating system.
+ "longName": "A String", # The operating system long name. For example 'Debian GNU/Linux 9' or 'Microsoft Window Server 2019 Datacenter'.
+ "osconfigAgentVersion": "A String", # The current version of the OS Config agent running on the VM.
+ "shortName": "A String", # The operating system short name. For example, 'windows' or 'debian'.
+ "version": "A String", # The version of the operating system.
+ },
+ "updateTime": "A String", # Output only. Timestamp of the last reported inventory for the VM.
+ },
+ ],
+ "nextPageToken": "A String", # The pagination token to retrieve the next page of inventory objects.
+}
+
+
+
+ list_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/osconfig_v1.projects.locations.instances.vulnerabilityReports.html b/docs/dyn/osconfig_v1.projects.locations.instances.vulnerabilityReports.html
new file mode 100644
index 00000000000..b4858dc3545
--- /dev/null
+++ b/docs/dyn/osconfig_v1.projects.locations.instances.vulnerabilityReports.html
@@ -0,0 +1,231 @@
+
+
+
+OS Config API . projects . locations . instances . vulnerabilityReports
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+Gets the vulnerability report for the specified VM instance. Only VMs with inventory data have vulnerability reports associated with them.
+
+ list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+List vulnerability reports for all VM instances in the specified zone.
+
+ list_next(previous_request, previous_response)
+Retrieves the next page of results.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ get(name, x__xgafv=None)
+ Gets the vulnerability report for the specified VM instance. Only VMs with inventory data have vulnerability reports associated with them.
+
+Args:
+ name: string, Required. API resource name for vulnerability resource. Format: `projects/{project}/locations/{location}/instances/{instance}/vulnerabilityReport` For `{project}`, either `project-number` or `project-id` can be provided. For `{instance}`, either Compute Engine `instance-id` or `instance-name` can be provided. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This API resource represents the vulnerability report for a specified Compute Engine virtual machine (VM) instance at a given point in time. For more information, see [Vulnerability reports](https://cloud.google.com/compute/docs/instances/os-inventory-management#vulnerability-reports).
+ "name": "A String", # Output only. The `vulnerabilityReport` API resource name. Format: `projects/{project_number}/locations/{location}/instances/{instance_id}/vulnerabilityReport`
+ "updateTime": "A String", # Output only. The timestamp for when the last vulnerability report was generated for the VM.
+ "vulnerabilities": [ # Output only. List of vulnerabilities affecting the VM.
+ { # A vulnerability affecting the VM instance.
+ "availableInventoryItemIds": [ # Corresponds to the `AVAILABLE_PACKAGE` inventory item on the VM. If the vulnerability report was not updated after the VM inventory update, these values might not display in VM inventory. If there is no available fix, the field is empty. The `inventory_item` value specifies the latest `SoftwarePackage` available to the VM that fixes the vulnerability.
+ "A String",
+ ],
+ "createTime": "A String", # The timestamp for when the vulnerability was first detected.
+ "details": { # Contains metadata information for the vulnerability. This information is collected from the upstream feed of the operating system. # Contains metadata as per the upstream feed of the operating system and NVD.
+ "cve": "A String", # The CVE of the vulnerability. CVE cannot be empty and the combination of should be unique across vulnerabilities for a VM.
+ "cvssV2Score": 3.14, # The CVSS V2 score of this vulnerability. CVSS V2 score is on a scale of 0 - 10 where 0 indicates low severity and 10 indicates high severity.
+ "cvssV3": { # Common Vulnerability Scoring System version 3. For details, see https://www.first.org/cvss/specification-document # The full description of the CVSSv3 for this vulnerability from NVD.
+ "attackComplexity": "A String", # This metric describes the conditions beyond the attacker's control that must exist in order to exploit the vulnerability.
+ "attackVector": "A String", # This metric reflects the context by which vulnerability exploitation is possible.
+ "availabilityImpact": "A String", # This metric measures the impact to the availability of the impacted component resulting from a successfully exploited vulnerability.
+ "baseScore": 3.14, # The base score is a function of the base metric scores. https://www.first.org/cvss/specification-document#Base-Metrics
+ "confidentialityImpact": "A String", # This metric measures the impact to the confidentiality of the information resources managed by a software component due to a successfully exploited vulnerability.
+ "exploitabilityScore": 3.14, # The Exploitability sub-score equation is derived from the Base Exploitability metrics. https://www.first.org/cvss/specification-document#2-1-Exploitability-Metrics
+ "impactScore": 3.14, # The Impact sub-score equation is derived from the Base Impact metrics.
+ "integrityImpact": "A String", # This metric measures the impact to integrity of a successfully exploited vulnerability.
+ "privilegesRequired": "A String", # This metric describes the level of privileges an attacker must possess before successfully exploiting the vulnerability.
+ "scope": "A String", # The Scope metric captures whether a vulnerability in one vulnerable component impacts resources in components beyond its security scope.
+ "userInteraction": "A String", # This metric captures the requirement for a human user, other than the attacker, to participate in the successful compromise of the vulnerable component.
+ },
+ "description": "A String", # The note or description describing the vulnerability from the distro.
+ "references": [ # Corresponds to the references attached to the `VulnerabilityDetails`.
+ { # A reference for this vulnerability.
+ "source": "A String", # The source of the reference e.g. NVD.
+ "url": "A String", # The url of the reference.
+ },
+ ],
+ "severity": "A String", # Assigned severity/impact ranking from the distro.
+ },
+ "installedInventoryItemIds": [ # Corresponds to the `INSTALLED_PACKAGE` inventory item on the VM. This field displays the inventory items affected by this vulnerability. If the vulnerability report was not updated after the VM inventory update, these values might not display in VM inventory. For some distros, this field may be empty.
+ "A String",
+ ],
+ "updateTime": "A String", # The timestamp for when the vulnerability was last modified.
+ },
+ ],
+}
+
+
+
+ list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+ List vulnerability reports for all VM instances in the specified zone.
+
+Args:
+ parent: string, Required. The parent resource name. Format: `projects/{project}/locations/{location}/instances/-` For `{project}`, either `project-number` or `project-id` can be provided. (required)
+ filter: string, If provided, this field specifies the criteria that must be met by a `vulnerabilityReport` API resource to be included in the response.
+ pageSize: integer, The maximum number of results to return.
+ pageToken: string, A pagination token returned from a previous call to `ListVulnerabilityReports` that indicates where this listing should continue from.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A response message for listing vulnerability reports for all VM instances in the specified location.
+ "nextPageToken": "A String", # The pagination token to retrieve the next page of vulnerabilityReports object.
+ "vulnerabilityReports": [ # List of vulnerabilityReport objects.
+ { # This API resource represents the vulnerability report for a specified Compute Engine virtual machine (VM) instance at a given point in time. For more information, see [Vulnerability reports](https://cloud.google.com/compute/docs/instances/os-inventory-management#vulnerability-reports).
+ "name": "A String", # Output only. The `vulnerabilityReport` API resource name. Format: `projects/{project_number}/locations/{location}/instances/{instance_id}/vulnerabilityReport`
+ "updateTime": "A String", # Output only. The timestamp for when the last vulnerability report was generated for the VM.
+ "vulnerabilities": [ # Output only. List of vulnerabilities affecting the VM.
+ { # A vulnerability affecting the VM instance.
+ "availableInventoryItemIds": [ # Corresponds to the `AVAILABLE_PACKAGE` inventory item on the VM. If the vulnerability report was not updated after the VM inventory update, these values might not display in VM inventory. If there is no available fix, the field is empty. The `inventory_item` value specifies the latest `SoftwarePackage` available to the VM that fixes the vulnerability.
+ "A String",
+ ],
+ "createTime": "A String", # The timestamp for when the vulnerability was first detected.
+ "details": { # Contains metadata information for the vulnerability. This information is collected from the upstream feed of the operating system. # Contains metadata as per the upstream feed of the operating system and NVD.
+ "cve": "A String", # The CVE of the vulnerability. CVE cannot be empty and the combination of should be unique across vulnerabilities for a VM.
+ "cvssV2Score": 3.14, # The CVSS V2 score of this vulnerability. CVSS V2 score is on a scale of 0 - 10 where 0 indicates low severity and 10 indicates high severity.
+ "cvssV3": { # Common Vulnerability Scoring System version 3. For details, see https://www.first.org/cvss/specification-document # The full description of the CVSSv3 for this vulnerability from NVD.
+ "attackComplexity": "A String", # This metric describes the conditions beyond the attacker's control that must exist in order to exploit the vulnerability.
+ "attackVector": "A String", # This metric reflects the context by which vulnerability exploitation is possible.
+ "availabilityImpact": "A String", # This metric measures the impact to the availability of the impacted component resulting from a successfully exploited vulnerability.
+ "baseScore": 3.14, # The base score is a function of the base metric scores. https://www.first.org/cvss/specification-document#Base-Metrics
+ "confidentialityImpact": "A String", # This metric measures the impact to the confidentiality of the information resources managed by a software component due to a successfully exploited vulnerability.
+ "exploitabilityScore": 3.14, # The Exploitability sub-score equation is derived from the Base Exploitability metrics. https://www.first.org/cvss/specification-document#2-1-Exploitability-Metrics
+ "impactScore": 3.14, # The Impact sub-score equation is derived from the Base Impact metrics.
+ "integrityImpact": "A String", # This metric measures the impact to integrity of a successfully exploited vulnerability.
+ "privilegesRequired": "A String", # This metric describes the level of privileges an attacker must possess before successfully exploiting the vulnerability.
+ "scope": "A String", # The Scope metric captures whether a vulnerability in one vulnerable component impacts resources in components beyond its security scope.
+ "userInteraction": "A String", # This metric captures the requirement for a human user, other than the attacker, to participate in the successful compromise of the vulnerable component.
+ },
+ "description": "A String", # The note or description describing the vulnerability from the distro.
+ "references": [ # Corresponds to the references attached to the `VulnerabilityDetails`.
+ { # A reference for this vulnerability.
+ "source": "A String", # The source of the reference e.g. NVD.
+ "url": "A String", # The url of the reference.
+ },
+ ],
+ "severity": "A String", # Assigned severity/impact ranking from the distro.
+ },
+ "installedInventoryItemIds": [ # Corresponds to the `INSTALLED_PACKAGE` inventory item on the VM. This field displays the inventory items affected by this vulnerability. If the vulnerability report was not updated after the VM inventory update, these values might not display in VM inventory. For some distros, this field may be empty.
+ "A String",
+ ],
+ "updateTime": "A String", # The timestamp for when the vulnerability was last modified.
+ },
+ ],
+ },
+ ],
+}
+
+
+
+ list_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/osconfig_v1alpha.projects.locations.instances.inventories.html b/docs/dyn/osconfig_v1alpha.projects.locations.instances.inventories.html
index 19398dfa40a..51fec74651d 100644
--- a/docs/dyn/osconfig_v1alpha.projects.locations.instances.inventories.html
+++ b/docs/dyn/osconfig_v1alpha.projects.locations.instances.inventories.html
@@ -136,6 +136,17 @@ Method Details
"hotFixId": "A String", # Unique identifier associated with a particular QFE update.
"installTime": "A String", # Date that the QFE update was installed. Mapped from installed_on field.
},
+ "windowsApplication": { # Contains information about a Windows application as retrieved from the Windows Registry. For more information about these fields, see [Windows Installer Properties for the Uninstall Registry](https://docs.microsoft.com/en-us/windows/win32/msi/uninstall-registry-key){: class="external" } # Details of Windows Application.
+ "displayName": "A String", # The name of the application or product.
+ "displayVersion": "A String", # The version of the product or application in string format.
+ "helpLink": "A String", # The internet address for technical support.
+ "installDate": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values * A month and day value, with a zero year, such as an anniversary * A year on its own, with zero month and day values * A year and month value, with a zero day, such as a credit card expiration date Related types are google.type.TimeOfDay and `google.protobuf.Timestamp`. # The last time this product received service. The value of this property is replaced each time a patch is applied or removed from the product or the command-line option is used to repair the product.
+ "day": 42, # Day of a month. Must be from 1 to 31 and valid for the year and month, or 0 to specify a year by itself or a year and month where the day isn't significant.
+ "month": 42, # Month of a year. Must be from 1 to 12, or 0 to specify a year without a month and day.
+ "year": 42, # Year of the date. Must be from 1 to 9999, or 0 to specify a date without a year.
+ },
+ "publisher": "A String", # The name of the manufacturer for the product or application.
+ },
"wuaPackage": { # Details related to a Windows Update package. Field data and names are taken from Windows Update API IUpdate Interface: https://docs.microsoft.com/en-us/windows/win32/api/_wua/ Descriptive fields like title, and description are localized based on the locale of the VM being updated. # Details of a Windows Update package. See https://docs.microsoft.com/en-us/windows/win32/api/_wua/ for information about Windows Update.
"categories": [ # The categories that are associated with this update package.
{ # Categories specified by the Windows Update.
@@ -197,6 +208,17 @@ Method Details
"hotFixId": "A String", # Unique identifier associated with a particular QFE update.
"installTime": "A String", # Date that the QFE update was installed. Mapped from installed_on field.
},
+ "windowsApplication": { # Contains information about a Windows application as retrieved from the Windows Registry. For more information about these fields, see [Windows Installer Properties for the Uninstall Registry](https://docs.microsoft.com/en-us/windows/win32/msi/uninstall-registry-key){: class="external" } # Details of Windows Application.
+ "displayName": "A String", # The name of the application or product.
+ "displayVersion": "A String", # The version of the product or application in string format.
+ "helpLink": "A String", # The internet address for technical support.
+ "installDate": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values * A month and day value, with a zero year, such as an anniversary * A year on its own, with zero month and day values * A year and month value, with a zero day, such as a credit card expiration date Related types are google.type.TimeOfDay and `google.protobuf.Timestamp`. # The last time this product received service. The value of this property is replaced each time a patch is applied or removed from the product or the command-line option is used to repair the product.
+ "day": 42, # Day of a month. Must be from 1 to 31 and valid for the year and month, or 0 to specify a year by itself or a year and month where the day isn't significant.
+ "month": 42, # Month of a year. Must be from 1 to 12, or 0 to specify a year without a month and day.
+ "year": 42, # Year of the date. Must be from 1 to 9999, or 0 to specify a date without a year.
+ },
+ "publisher": "A String", # The name of the manufacturer for the product or application.
+ },
"wuaPackage": { # Details related to a Windows Update package. Field data and names are taken from Windows Update API IUpdate Interface: https://docs.microsoft.com/en-us/windows/win32/api/_wua/ Descriptive fields like title, and description are localized based on the locale of the VM being updated. # Details of a Windows Update package. See https://docs.microsoft.com/en-us/windows/win32/api/_wua/ for information about Windows Update.
"categories": [ # The categories that are associated with this update package.
{ # Categories specified by the Windows Update.
@@ -259,7 +281,7 @@ Method Details
List inventory data for all VM instances in the specified zone.
Args:
- parent: string, Required. The parent resource name. Format: `projects/{project}/locations/{location}/instances/{instance}` For `{project}`, either `project-number` or `project-id` can be provided. For `{instance}`, only hyphen or dash character is supported to list inventories across VMs. (required)
+ parent: string, Required. The parent resource name. Format: `projects/{project}/locations/{location}/instances/-` For `{project}`, either `project-number` or `project-id` can be provided. (required)
filter: string, If provided, this field specifies the criteria that must be met by a `Inventory` API resource to be included in the response.
pageSize: integer, The maximum number of results to return.
pageToken: string, A pagination token returned from a previous call to `ListInventories` that indicates where this listing should continue from.
@@ -303,6 +325,17 @@ Method Details
"hotFixId": "A String", # Unique identifier associated with a particular QFE update.
"installTime": "A String", # Date that the QFE update was installed. Mapped from installed_on field.
},
+ "windowsApplication": { # Contains information about a Windows application as retrieved from the Windows Registry. For more information about these fields, see [Windows Installer Properties for the Uninstall Registry](https://docs.microsoft.com/en-us/windows/win32/msi/uninstall-registry-key){: class="external" } # Details of Windows Application.
+ "displayName": "A String", # The name of the application or product.
+ "displayVersion": "A String", # The version of the product or application in string format.
+ "helpLink": "A String", # The internet address for technical support.
+ "installDate": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values * A month and day value, with a zero year, such as an anniversary * A year on its own, with zero month and day values * A year and month value, with a zero day, such as a credit card expiration date Related types are google.type.TimeOfDay and `google.protobuf.Timestamp`. # The last time this product received service. The value of this property is replaced each time a patch is applied or removed from the product or the command-line option is used to repair the product.
+ "day": 42, # Day of a month. Must be from 1 to 31 and valid for the year and month, or 0 to specify a year by itself or a year and month where the day isn't significant.
+ "month": 42, # Month of a year. Must be from 1 to 12, or 0 to specify a year without a month and day.
+ "year": 42, # Year of the date. Must be from 1 to 9999, or 0 to specify a date without a year.
+ },
+ "publisher": "A String", # The name of the manufacturer for the product or application.
+ },
"wuaPackage": { # Details related to a Windows Update package. Field data and names are taken from Windows Update API IUpdate Interface: https://docs.microsoft.com/en-us/windows/win32/api/_wua/ Descriptive fields like title, and description are localized based on the locale of the VM being updated. # Details of a Windows Update package. See https://docs.microsoft.com/en-us/windows/win32/api/_wua/ for information about Windows Update.
"categories": [ # The categories that are associated with this update package.
{ # Categories specified by the Windows Update.
@@ -364,6 +397,17 @@ Method Details
"hotFixId": "A String", # Unique identifier associated with a particular QFE update.
"installTime": "A String", # Date that the QFE update was installed. Mapped from installed_on field.
},
+ "windowsApplication": { # Contains information about a Windows application as retrieved from the Windows Registry. For more information about these fields, see [Windows Installer Properties for the Uninstall Registry](https://docs.microsoft.com/en-us/windows/win32/msi/uninstall-registry-key){: class="external" } # Details of Windows Application.
+ "displayName": "A String", # The name of the application or product.
+ "displayVersion": "A String", # The version of the product or application in string format.
+ "helpLink": "A String", # The internet address for technical support.
+ "installDate": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values * A month and day value, with a zero year, such as an anniversary * A year on its own, with zero month and day values * A year and month value, with a zero day, such as a credit card expiration date Related types are google.type.TimeOfDay and `google.protobuf.Timestamp`. # The last time this product received service. The value of this property is replaced each time a patch is applied or removed from the product or the command-line option is used to repair the product.
+ "day": 42, # Day of a month. Must be from 1 to 31 and valid for the year and month, or 0 to specify a year by itself or a year and month where the day isn't significant.
+ "month": 42, # Month of a year. Must be from 1 to 12, or 0 to specify a year without a month and day.
+ "year": 42, # Year of the date. Must be from 1 to 9999, or 0 to specify a date without a year.
+ },
+ "publisher": "A String", # The name of the manufacturer for the product or application.
+ },
"wuaPackage": { # Details related to a Windows Update package. Field data and names are taken from Windows Update API IUpdate Interface: https://docs.microsoft.com/en-us/windows/win32/api/_wua/ Descriptive fields like title, and description are localized based on the locale of the VM being updated. # Details of a Windows Update package. See https://docs.microsoft.com/en-us/windows/win32/api/_wua/ for information about Windows Update.
"categories": [ # The categories that are associated with this update package.
{ # Categories specified by the Windows Update.
diff --git a/docs/dyn/osconfig_v1alpha.projects.locations.instances.vulnerabilityReports.html b/docs/dyn/osconfig_v1alpha.projects.locations.instances.vulnerabilityReports.html
index 2547a77654f..75ddff24a47 100644
--- a/docs/dyn/osconfig_v1alpha.projects.locations.instances.vulnerabilityReports.html
+++ b/docs/dyn/osconfig_v1alpha.projects.locations.instances.vulnerabilityReports.html
@@ -134,6 +134,7 @@ Method Details
"description": "A String", # The note or description describing the vulnerability from the distro.
"references": [ # Corresponds to the references attached to the `VulnerabilityDetails`.
{ # A reference for this vulnerability.
+ "source": "A String", # The source of the reference e.g. NVD.
"url": "A String", # The url of the reference.
},
],
@@ -153,7 +154,7 @@ Method Details
List vulnerability reports for all VM instances in the specified zone.
Args:
- parent: string, Required. The parent resource name. Format: `projects/{project}/locations/{location}/instances/{instance}` For `{project}`, either `project-number` or `project-id` can be provided. For `{instance}`, only `-` character is supported to list vulnerability reports across VMs. (required)
+ parent: string, Required. The parent resource name. Format: `projects/{project}/locations/{location}/instances/-` For `{project}`, either `project-number` or `project-id` can be provided. (required)
filter: string, If provided, this field specifies the criteria that must be met by a `vulnerabilityReport` API resource to be included in the response.
pageSize: integer, The maximum number of results to return.
pageToken: string, A pagination token returned from a previous call to `ListVulnerabilityReports` that indicates where this listing should continue from.
@@ -196,6 +197,7 @@ Method Details
"description": "A String", # The note or description describing the vulnerability from the distro.
"references": [ # Corresponds to the references attached to the `VulnerabilityDetails`.
{ # A reference for this vulnerability.
+ "source": "A String", # The source of the reference e.g. NVD.
"url": "A String", # The url of the reference.
},
],
diff --git a/docs/dyn/osconfig_v1alpha.projects.locations.osPolicyAssignments.html b/docs/dyn/osconfig_v1alpha.projects.locations.osPolicyAssignments.html
index b5091cd0e60..7518189b60d 100644
--- a/docs/dyn/osconfig_v1alpha.projects.locations.osPolicyAssignments.html
+++ b/docs/dyn/osconfig_v1alpha.projects.locations.osPolicyAssignments.html
@@ -125,9 +125,10 @@ Method Details
"baseline": True or False, # Output only. Indicates that this revision has been successfully rolled out in this zone and new VMs will be assigned OS policies from this revision. For a given OS policy assignment, there is only one revision with a value of `true` for this field.
"deleted": True or False, # Output only. Indicates that this revision deletes the OS policy assignment.
"description": "A String", # OS policy assignment description. Length of the description is limited to 1024 characters.
- "instanceFilter": { # Message to represent the filters to select VMs for an assignment # Required. Filter to select VMs.
+ "etag": "A String", # The etag for this OS policy assignment. If this is provided on update, it must match the server's etag.
+ "instanceFilter": { # Filters to select target VMs for an assignment. If more than one filter criteria is specified below, a VM will be selected if and only if it satisfies all of them. # Required. Filter to select VMs.
"all": True or False, # Target all VMs in the project. If true, no other criteria is permitted.
- "exclusionLabels": [ # List of label sets used for VM exclusion. If the list has more than one label set, the VM is excluded if any of the label sets are applicable for the VM. This filter is applied last in the filtering chain and therefore a VM is guaranteed to be excluded if it satisfies one of the below label sets.
+ "exclusionLabels": [ # List of label sets used for VM exclusion. If the list has more than one label set, the VM is excluded if any of the label sets are applicable for the VM.
{ # Message representing label set. * A label is a key value pair set for a VM. * A LabelSet is a set of labels. * Labels within a LabelSet are ANDed. In other words, a LabelSet is applicable for a VM only if it matches all the labels in the LabelSet. * Example: A LabelSet with 2 labels: `env=prod` and `type=webserver` will only be applicable for those VMs with both labels present.
"labels": { # Labels are identified by key/value pairs in this map. A VM should contain all the key/value pairs specified in this map to be selected.
"a_key": "A String",
@@ -141,7 +142,13 @@ Method Details
},
},
],
- "osShortNames": [ # A VM is included if it's OS short name matches with any of the values provided in this list.
+ "inventories": [ # List of inventories to select VMs. A VM is selected if its inventory data matches at least one of the following inventories.
+ { # VM inventory details.
+ "osShortName": "A String", # Required. The OS short name
+ "osVersion": "A String", # The OS version Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*` An empty string matches all OS versions.
+ },
+ ],
+ "osShortNames": [ # A VM is selected if it's OS short name matches with any of the values provided in this list.
"A String",
],
},
@@ -154,7 +161,13 @@ Method Details
"mode": "A String", # Required. Policy mode
"resourceGroups": [ # Required. List of resource groups for the policy. For a particular VM, resource groups are evaluated in the order specified and the first resource group that is applicable is selected and the rest are ignored. If none of the resource groups are applicable for a VM, the VM is considered to be non-compliant w.r.t this policy. This behavior can be toggled by the flag `allow_no_resource_group_match`
{ # Resource groups provide a mechanism to group OS policy resources. Resource groups enable OS policy authors to create a single OS policy to be applied to VMs running different operating Systems. When the OS policy is applied to a target VM, the appropriate resource group within the OS policy is selected based on the `OSFilter` specified within the resource group.
- "osFilter": { # The `OSFilter` is used to specify the OS filtering criteria for the resource group. # Used to specify the OS filter for a resource group
+ "inventoryFilters": [ # List of inventory filters for the resource group. The resources in this resource group are applied to the target VM if it satisfies at least one of the following inventory filters. For example, to apply this resource group to VMs running either `RHEL` or `CentOS` operating systems, specify 2 items for the list with following values: inventory_filters[0].os_short_name='rhel' and inventory_filters[1].os_short_name='centos' If the list is empty, this resource group will be applied to the target VM unconditionally.
+ { # Filtering criteria to select VMs based on inventory details.
+ "osShortName": "A String", # Required. The OS short name
+ "osVersion": "A String", # The OS version Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*` An empty string matches all OS versions.
+ },
+ ],
+ "osFilter": { # Filtering criteria to select VMs based on OS details. # Used to specify the OS filter for a resource group
"osShortName": "A String", # This should match OS short name emitted by the OS inventory agent. An empty value matches any OS.
"osVersion": "A String", # This value should match the version emitted by the OS inventory agent. Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*`
},
@@ -423,9 +436,10 @@ Method Details
"baseline": True or False, # Output only. Indicates that this revision has been successfully rolled out in this zone and new VMs will be assigned OS policies from this revision. For a given OS policy assignment, there is only one revision with a value of `true` for this field.
"deleted": True or False, # Output only. Indicates that this revision deletes the OS policy assignment.
"description": "A String", # OS policy assignment description. Length of the description is limited to 1024 characters.
- "instanceFilter": { # Message to represent the filters to select VMs for an assignment # Required. Filter to select VMs.
+ "etag": "A String", # The etag for this OS policy assignment. If this is provided on update, it must match the server's etag.
+ "instanceFilter": { # Filters to select target VMs for an assignment. If more than one filter criteria is specified below, a VM will be selected if and only if it satisfies all of them. # Required. Filter to select VMs.
"all": True or False, # Target all VMs in the project. If true, no other criteria is permitted.
- "exclusionLabels": [ # List of label sets used for VM exclusion. If the list has more than one label set, the VM is excluded if any of the label sets are applicable for the VM. This filter is applied last in the filtering chain and therefore a VM is guaranteed to be excluded if it satisfies one of the below label sets.
+ "exclusionLabels": [ # List of label sets used for VM exclusion. If the list has more than one label set, the VM is excluded if any of the label sets are applicable for the VM.
{ # Message representing label set. * A label is a key value pair set for a VM. * A LabelSet is a set of labels. * Labels within a LabelSet are ANDed. In other words, a LabelSet is applicable for a VM only if it matches all the labels in the LabelSet. * Example: A LabelSet with 2 labels: `env=prod` and `type=webserver` will only be applicable for those VMs with both labels present.
"labels": { # Labels are identified by key/value pairs in this map. A VM should contain all the key/value pairs specified in this map to be selected.
"a_key": "A String",
@@ -439,7 +453,13 @@ Method Details
},
},
],
- "osShortNames": [ # A VM is included if it's OS short name matches with any of the values provided in this list.
+ "inventories": [ # List of inventories to select VMs. A VM is selected if its inventory data matches at least one of the following inventories.
+ { # VM inventory details.
+ "osShortName": "A String", # Required. The OS short name
+ "osVersion": "A String", # The OS version Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*` An empty string matches all OS versions.
+ },
+ ],
+ "osShortNames": [ # A VM is selected if it's OS short name matches with any of the values provided in this list.
"A String",
],
},
@@ -452,7 +472,13 @@ Method Details
"mode": "A String", # Required. Policy mode
"resourceGroups": [ # Required. List of resource groups for the policy. For a particular VM, resource groups are evaluated in the order specified and the first resource group that is applicable is selected and the rest are ignored. If none of the resource groups are applicable for a VM, the VM is considered to be non-compliant w.r.t this policy. This behavior can be toggled by the flag `allow_no_resource_group_match`
{ # Resource groups provide a mechanism to group OS policy resources. Resource groups enable OS policy authors to create a single OS policy to be applied to VMs running different operating Systems. When the OS policy is applied to a target VM, the appropriate resource group within the OS policy is selected based on the `OSFilter` specified within the resource group.
- "osFilter": { # The `OSFilter` is used to specify the OS filtering criteria for the resource group. # Used to specify the OS filter for a resource group
+ "inventoryFilters": [ # List of inventory filters for the resource group. The resources in this resource group are applied to the target VM if it satisfies at least one of the following inventory filters. For example, to apply this resource group to VMs running either `RHEL` or `CentOS` operating systems, specify 2 items for the list with following values: inventory_filters[0].os_short_name='rhel' and inventory_filters[1].os_short_name='centos' If the list is empty, this resource group will be applied to the target VM unconditionally.
+ { # Filtering criteria to select VMs based on inventory details.
+ "osShortName": "A String", # Required. The OS short name
+ "osVersion": "A String", # The OS version Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*` An empty string matches all OS versions.
+ },
+ ],
+ "osFilter": { # Filtering criteria to select VMs based on OS details. # Used to specify the OS filter for a resource group
"osShortName": "A String", # This should match OS short name emitted by the OS inventory agent. An empty value matches any OS.
"osVersion": "A String", # This value should match the version emitted by the OS inventory agent. Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*`
},
@@ -662,9 +688,10 @@ Method Details
"baseline": True or False, # Output only. Indicates that this revision has been successfully rolled out in this zone and new VMs will be assigned OS policies from this revision. For a given OS policy assignment, there is only one revision with a value of `true` for this field.
"deleted": True or False, # Output only. Indicates that this revision deletes the OS policy assignment.
"description": "A String", # OS policy assignment description. Length of the description is limited to 1024 characters.
- "instanceFilter": { # Message to represent the filters to select VMs for an assignment # Required. Filter to select VMs.
+ "etag": "A String", # The etag for this OS policy assignment. If this is provided on update, it must match the server's etag.
+ "instanceFilter": { # Filters to select target VMs for an assignment. If more than one filter criteria is specified below, a VM will be selected if and only if it satisfies all of them. # Required. Filter to select VMs.
"all": True or False, # Target all VMs in the project. If true, no other criteria is permitted.
- "exclusionLabels": [ # List of label sets used for VM exclusion. If the list has more than one label set, the VM is excluded if any of the label sets are applicable for the VM. This filter is applied last in the filtering chain and therefore a VM is guaranteed to be excluded if it satisfies one of the below label sets.
+ "exclusionLabels": [ # List of label sets used for VM exclusion. If the list has more than one label set, the VM is excluded if any of the label sets are applicable for the VM.
{ # Message representing label set. * A label is a key value pair set for a VM. * A LabelSet is a set of labels. * Labels within a LabelSet are ANDed. In other words, a LabelSet is applicable for a VM only if it matches all the labels in the LabelSet. * Example: A LabelSet with 2 labels: `env=prod` and `type=webserver` will only be applicable for those VMs with both labels present.
"labels": { # Labels are identified by key/value pairs in this map. A VM should contain all the key/value pairs specified in this map to be selected.
"a_key": "A String",
@@ -678,7 +705,13 @@ Method Details
},
},
],
- "osShortNames": [ # A VM is included if it's OS short name matches with any of the values provided in this list.
+ "inventories": [ # List of inventories to select VMs. A VM is selected if its inventory data matches at least one of the following inventories.
+ { # VM inventory details.
+ "osShortName": "A String", # Required. The OS short name
+ "osVersion": "A String", # The OS version Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*` An empty string matches all OS versions.
+ },
+ ],
+ "osShortNames": [ # A VM is selected if it's OS short name matches with any of the values provided in this list.
"A String",
],
},
@@ -691,7 +724,13 @@ Method Details
"mode": "A String", # Required. Policy mode
"resourceGroups": [ # Required. List of resource groups for the policy. For a particular VM, resource groups are evaluated in the order specified and the first resource group that is applicable is selected and the rest are ignored. If none of the resource groups are applicable for a VM, the VM is considered to be non-compliant w.r.t this policy. This behavior can be toggled by the flag `allow_no_resource_group_match`
{ # Resource groups provide a mechanism to group OS policy resources. Resource groups enable OS policy authors to create a single OS policy to be applied to VMs running different operating Systems. When the OS policy is applied to a target VM, the appropriate resource group within the OS policy is selected based on the `OSFilter` specified within the resource group.
- "osFilter": { # The `OSFilter` is used to specify the OS filtering criteria for the resource group. # Used to specify the OS filter for a resource group
+ "inventoryFilters": [ # List of inventory filters for the resource group. The resources in this resource group are applied to the target VM if it satisfies at least one of the following inventory filters. For example, to apply this resource group to VMs running either `RHEL` or `CentOS` operating systems, specify 2 items for the list with following values: inventory_filters[0].os_short_name='rhel' and inventory_filters[1].os_short_name='centos' If the list is empty, this resource group will be applied to the target VM unconditionally.
+ { # Filtering criteria to select VMs based on inventory details.
+ "osShortName": "A String", # Required. The OS short name
+ "osVersion": "A String", # The OS version Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*` An empty string matches all OS versions.
+ },
+ ],
+ "osFilter": { # Filtering criteria to select VMs based on OS details. # Used to specify the OS filter for a resource group
"osShortName": "A String", # This should match OS short name emitted by the OS inventory agent. An empty value matches any OS.
"osVersion": "A String", # This value should match the version emitted by the OS inventory agent. Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*`
},
@@ -903,9 +942,10 @@ Method Details
"baseline": True or False, # Output only. Indicates that this revision has been successfully rolled out in this zone and new VMs will be assigned OS policies from this revision. For a given OS policy assignment, there is only one revision with a value of `true` for this field.
"deleted": True or False, # Output only. Indicates that this revision deletes the OS policy assignment.
"description": "A String", # OS policy assignment description. Length of the description is limited to 1024 characters.
- "instanceFilter": { # Message to represent the filters to select VMs for an assignment # Required. Filter to select VMs.
+ "etag": "A String", # The etag for this OS policy assignment. If this is provided on update, it must match the server's etag.
+ "instanceFilter": { # Filters to select target VMs for an assignment. If more than one filter criteria is specified below, a VM will be selected if and only if it satisfies all of them. # Required. Filter to select VMs.
"all": True or False, # Target all VMs in the project. If true, no other criteria is permitted.
- "exclusionLabels": [ # List of label sets used for VM exclusion. If the list has more than one label set, the VM is excluded if any of the label sets are applicable for the VM. This filter is applied last in the filtering chain and therefore a VM is guaranteed to be excluded if it satisfies one of the below label sets.
+ "exclusionLabels": [ # List of label sets used for VM exclusion. If the list has more than one label set, the VM is excluded if any of the label sets are applicable for the VM.
{ # Message representing label set. * A label is a key value pair set for a VM. * A LabelSet is a set of labels. * Labels within a LabelSet are ANDed. In other words, a LabelSet is applicable for a VM only if it matches all the labels in the LabelSet. * Example: A LabelSet with 2 labels: `env=prod` and `type=webserver` will only be applicable for those VMs with both labels present.
"labels": { # Labels are identified by key/value pairs in this map. A VM should contain all the key/value pairs specified in this map to be selected.
"a_key": "A String",
@@ -919,7 +959,13 @@ Method Details
},
},
],
- "osShortNames": [ # A VM is included if it's OS short name matches with any of the values provided in this list.
+ "inventories": [ # List of inventories to select VMs. A VM is selected if its inventory data matches at least one of the following inventories.
+ { # VM inventory details.
+ "osShortName": "A String", # Required. The OS short name
+ "osVersion": "A String", # The OS version Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*` An empty string matches all OS versions.
+ },
+ ],
+ "osShortNames": [ # A VM is selected if it's OS short name matches with any of the values provided in this list.
"A String",
],
},
@@ -932,7 +978,13 @@ Method Details
"mode": "A String", # Required. Policy mode
"resourceGroups": [ # Required. List of resource groups for the policy. For a particular VM, resource groups are evaluated in the order specified and the first resource group that is applicable is selected and the rest are ignored. If none of the resource groups are applicable for a VM, the VM is considered to be non-compliant w.r.t this policy. This behavior can be toggled by the flag `allow_no_resource_group_match`
{ # Resource groups provide a mechanism to group OS policy resources. Resource groups enable OS policy authors to create a single OS policy to be applied to VMs running different operating Systems. When the OS policy is applied to a target VM, the appropriate resource group within the OS policy is selected based on the `OSFilter` specified within the resource group.
- "osFilter": { # The `OSFilter` is used to specify the OS filtering criteria for the resource group. # Used to specify the OS filter for a resource group
+ "inventoryFilters": [ # List of inventory filters for the resource group. The resources in this resource group are applied to the target VM if it satisfies at least one of the following inventory filters. For example, to apply this resource group to VMs running either `RHEL` or `CentOS` operating systems, specify 2 items for the list with following values: inventory_filters[0].os_short_name='rhel' and inventory_filters[1].os_short_name='centos' If the list is empty, this resource group will be applied to the target VM unconditionally.
+ { # Filtering criteria to select VMs based on inventory details.
+ "osShortName": "A String", # Required. The OS short name
+ "osVersion": "A String", # The OS version Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*` An empty string matches all OS versions.
+ },
+ ],
+ "osFilter": { # Filtering criteria to select VMs based on OS details. # Used to specify the OS filter for a resource group
"osShortName": "A String", # This should match OS short name emitted by the OS inventory agent. An empty value matches any OS.
"osVersion": "A String", # This value should match the version emitted by the OS inventory agent. Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*`
},
@@ -1162,9 +1214,10 @@ Method Details
"baseline": True or False, # Output only. Indicates that this revision has been successfully rolled out in this zone and new VMs will be assigned OS policies from this revision. For a given OS policy assignment, there is only one revision with a value of `true` for this field.
"deleted": True or False, # Output only. Indicates that this revision deletes the OS policy assignment.
"description": "A String", # OS policy assignment description. Length of the description is limited to 1024 characters.
- "instanceFilter": { # Message to represent the filters to select VMs for an assignment # Required. Filter to select VMs.
+ "etag": "A String", # The etag for this OS policy assignment. If this is provided on update, it must match the server's etag.
+ "instanceFilter": { # Filters to select target VMs for an assignment. If more than one filter criteria is specified below, a VM will be selected if and only if it satisfies all of them. # Required. Filter to select VMs.
"all": True or False, # Target all VMs in the project. If true, no other criteria is permitted.
- "exclusionLabels": [ # List of label sets used for VM exclusion. If the list has more than one label set, the VM is excluded if any of the label sets are applicable for the VM. This filter is applied last in the filtering chain and therefore a VM is guaranteed to be excluded if it satisfies one of the below label sets.
+ "exclusionLabels": [ # List of label sets used for VM exclusion. If the list has more than one label set, the VM is excluded if any of the label sets are applicable for the VM.
{ # Message representing label set. * A label is a key value pair set for a VM. * A LabelSet is a set of labels. * Labels within a LabelSet are ANDed. In other words, a LabelSet is applicable for a VM only if it matches all the labels in the LabelSet. * Example: A LabelSet with 2 labels: `env=prod` and `type=webserver` will only be applicable for those VMs with both labels present.
"labels": { # Labels are identified by key/value pairs in this map. A VM should contain all the key/value pairs specified in this map to be selected.
"a_key": "A String",
@@ -1178,7 +1231,13 @@ Method Details
},
},
],
- "osShortNames": [ # A VM is included if it's OS short name matches with any of the values provided in this list.
+ "inventories": [ # List of inventories to select VMs. A VM is selected if its inventory data matches at least one of the following inventories.
+ { # VM inventory details.
+ "osShortName": "A String", # Required. The OS short name
+ "osVersion": "A String", # The OS version Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*` An empty string matches all OS versions.
+ },
+ ],
+ "osShortNames": [ # A VM is selected if it's OS short name matches with any of the values provided in this list.
"A String",
],
},
@@ -1191,7 +1250,13 @@ Method Details
"mode": "A String", # Required. Policy mode
"resourceGroups": [ # Required. List of resource groups for the policy. For a particular VM, resource groups are evaluated in the order specified and the first resource group that is applicable is selected and the rest are ignored. If none of the resource groups are applicable for a VM, the VM is considered to be non-compliant w.r.t this policy. This behavior can be toggled by the flag `allow_no_resource_group_match`
{ # Resource groups provide a mechanism to group OS policy resources. Resource groups enable OS policy authors to create a single OS policy to be applied to VMs running different operating Systems. When the OS policy is applied to a target VM, the appropriate resource group within the OS policy is selected based on the `OSFilter` specified within the resource group.
- "osFilter": { # The `OSFilter` is used to specify the OS filtering criteria for the resource group. # Used to specify the OS filter for a resource group
+ "inventoryFilters": [ # List of inventory filters for the resource group. The resources in this resource group are applied to the target VM if it satisfies at least one of the following inventory filters. For example, to apply this resource group to VMs running either `RHEL` or `CentOS` operating systems, specify 2 items for the list with following values: inventory_filters[0].os_short_name='rhel' and inventory_filters[1].os_short_name='centos' If the list is empty, this resource group will be applied to the target VM unconditionally.
+ { # Filtering criteria to select VMs based on inventory details.
+ "osShortName": "A String", # Required. The OS short name
+ "osVersion": "A String", # The OS version Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*` An empty string matches all OS versions.
+ },
+ ],
+ "osFilter": { # Filtering criteria to select VMs based on OS details. # Used to specify the OS filter for a resource group
"osShortName": "A String", # This should match OS short name emitted by the OS inventory agent. An empty value matches any OS.
"osVersion": "A String", # This value should match the version emitted by the OS inventory agent. Prefix matches are supported if asterisk(*) is provided as the last character. For example, to match all versions with a major version of `7`, specify the following value for this field `7.*`
},
diff --git a/docs/dyn/oslogin_v1alpha.users.html b/docs/dyn/oslogin_v1alpha.users.html
index 14c8aaf5a27..3d5d134a3d8 100644
--- a/docs/dyn/oslogin_v1alpha.users.html
+++ b/docs/dyn/oslogin_v1alpha.users.html
@@ -142,6 +142,18 @@ Method Details
"username": "A String", # The username of the POSIX account.
},
],
+ "securityKeys": [ # The registered security key credentials for a user.
+ { # The credential information for a Google registered security key.
+ "privateKey": "A String", # Hardware-backed private key text in SSH format.
+ "publicKey": "A String", # Public key text in SSH format, defined by [RFC4253]("https://www.ietf.org/rfc/rfc4253.txt") section 6.6.
+ "universalTwoFactor": { # Security key information specific to the U2F protocol. # The U2F protocol type.
+ "appId": "A String", # Application ID for the U2F protocol.
+ },
+ "webAuthn": { # Security key information specific to the Web Authentication protocol. # The Web Authentication protocol type.
+ "rpId": "A String", # Relying party ID for Web Authentication.
+ },
+ },
+ ],
"sshPublicKeys": { # A map from SSH public key fingerprint to the associated key object.
"a_key": { # The SSH public key information associated with a Google account.
"expirationTimeUsec": "A String", # An expiration time in microseconds since epoch.
@@ -202,6 +214,18 @@ Method Details
"username": "A String", # The username of the POSIX account.
},
],
+ "securityKeys": [ # The registered security key credentials for a user.
+ { # The credential information for a Google registered security key.
+ "privateKey": "A String", # Hardware-backed private key text in SSH format.
+ "publicKey": "A String", # Public key text in SSH format, defined by [RFC4253]("https://www.ietf.org/rfc/rfc4253.txt") section 6.6.
+ "universalTwoFactor": { # Security key information specific to the U2F protocol. # The U2F protocol type.
+ "appId": "A String", # Application ID for the U2F protocol.
+ },
+ "webAuthn": { # Security key information specific to the Web Authentication protocol. # The Web Authentication protocol type.
+ "rpId": "A String", # Relying party ID for Web Authentication.
+ },
+ },
+ ],
"sshPublicKeys": { # A map from SSH public key fingerprint to the associated key object.
"a_key": { # The SSH public key information associated with a Google account.
"expirationTimeUsec": "A String", # An expiration time in microseconds since epoch.
diff --git a/docs/dyn/oslogin_v1beta.users.html b/docs/dyn/oslogin_v1beta.users.html
index 98a426d02d5..b26e6edb5cd 100644
--- a/docs/dyn/oslogin_v1beta.users.html
+++ b/docs/dyn/oslogin_v1beta.users.html
@@ -137,6 +137,18 @@ Method Details
"username": "A String", # The username of the POSIX account.
},
],
+ "securityKeys": [ # The registered security key credentials for a user.
+ { # The credential information for a Google registered security key.
+ "privateKey": "A String", # Hardware-backed private key text in SSH format.
+ "publicKey": "A String", # Public key text in SSH format, defined by [RFC4253]("https://www.ietf.org/rfc/rfc4253.txt") section 6.6.
+ "universalTwoFactor": { # Security key information specific to the U2F protocol. # The U2F protocol type.
+ "appId": "A String", # Application ID for the U2F protocol.
+ },
+ "webAuthn": { # Security key information specific to the Web Authentication protocol. # The Web Authentication protocol type.
+ "rpId": "A String", # Relying party ID for Web Authentication.
+ },
+ },
+ ],
"sshPublicKeys": { # A map from SSH public key fingerprint to the associated key object.
"a_key": { # The SSH public key information associated with a Google account.
"expirationTimeUsec": "A String", # An expiration time in microseconds since epoch.
@@ -197,6 +209,18 @@ Method Details
"username": "A String", # The username of the POSIX account.
},
],
+ "securityKeys": [ # The registered security key credentials for a user.
+ { # The credential information for a Google registered security key.
+ "privateKey": "A String", # Hardware-backed private key text in SSH format.
+ "publicKey": "A String", # Public key text in SSH format, defined by [RFC4253]("https://www.ietf.org/rfc/rfc4253.txt") section 6.6.
+ "universalTwoFactor": { # Security key information specific to the U2F protocol. # The U2F protocol type.
+ "appId": "A String", # Application ID for the U2F protocol.
+ },
+ "webAuthn": { # Security key information specific to the Web Authentication protocol. # The Web Authentication protocol type.
+ "rpId": "A String", # Relying party ID for Web Authentication.
+ },
+ },
+ ],
"sshPublicKeys": { # A map from SSH public key fingerprint to the associated key object.
"a_key": { # The SSH public key information associated with a Google account.
"expirationTimeUsec": "A String", # An expiration time in microseconds since epoch.
diff --git a/docs/dyn/people_v1.otherContacts.html b/docs/dyn/people_v1.otherContacts.html
index c6c31f8a219..a2b7963e473 100644
--- a/docs/dyn/people_v1.otherContacts.html
+++ b/docs/dyn/people_v1.otherContacts.html
@@ -1005,7 +1005,7 @@ Method Details
Args:
pageSize: integer, Optional. The number of "Other contacts" to include in the response. Valid values are between 1 and 1000, inclusive. Defaults to 100 if not set or set to 0.
pageToken: string, Optional. A page token, received from a previous response `next_page_token`. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `otherContacts.list` must match the first call that provided the page token.
- readMask: string, Required. A field mask to restrict which fields on each person are returned. Multiple fields can be specified by separating them with commas. Valid values are: * emailAddresses * metadata * names * phoneNumbers * photos
+ readMask: string, Required. A field mask to restrict which fields on each person are returned. Multiple fields can be specified by separating them with commas. What values are valid depend on what ReadSourceType is used. If READ_SOURCE_TYPE_CONTACT is used, valid values are: * emailAddresses * metadata * names * phoneNumbers * photos If READ_SOURCE_TYPE_PROFILE is used, valid values are: * addresses * ageRanges * biographies * birthdays * calendarUrls * clientData * coverPhotos * emailAddresses * events * externalIds * genders * imClients * interests * locales * locations * memberships * metadata * miscKeywords * names * nicknames * occupations * organizations * phoneNumbers * photos * relations * sipAddresses * skills * urls * userDefined
requestSyncToken: boolean, Optional. Whether the response should return `next_sync_token` on the last page of results. It can be used to get incremental changes since the last request by setting it on the request `sync_token`. More details about sync behavior at `otherContacts.list`.
sources: string, Optional. A mask of what source types to return. Defaults to READ_SOURCE_TYPE_CONTACT if not set. (repeated)
Allowed values
diff --git a/docs/dyn/policyanalyzer_v1.html b/docs/dyn/policyanalyzer_v1.html
new file mode 100644
index 00000000000..7d78e03c456
--- /dev/null
+++ b/docs/dyn/policyanalyzer_v1.html
@@ -0,0 +1,111 @@
+
+
+
+Policy Analyzer API
+Instance Methods
+
+ projects()
+
+Returns the projects Resource.
+
+
+ close()
+Close httplib2 connections.
+
+Create a BatchHttpRequest object based on the discovery document.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ new_batch_http_request()
+ Create a BatchHttpRequest object based on the discovery document.
+
+ Args:
+ callback: callable, A callback to be called for each response, of the
+ form callback(id, response, exception). The first parameter is the
+ request id, and the second is the deserialized response object. The
+ third is an apiclient.errors.HttpError exception object if an HTTP
+ error occurred while processing the request, or None if no error
+ occurred.
+
+ Returns:
+ A BatchHttpRequest object based on the discovery document.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/policyanalyzer_v1.projects.html b/docs/dyn/policyanalyzer_v1.projects.html
new file mode 100644
index 00000000000..d559943e654
--- /dev/null
+++ b/docs/dyn/policyanalyzer_v1.projects.html
@@ -0,0 +1,91 @@
+
+
+
+Policy Analyzer API . projects
+Instance Methods
+
+ locations()
+
+Returns the locations Resource.
+
+
+ close()
+Close httplib2 connections.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/policyanalyzer_v1.projects.locations.activityTypes.activities.html b/docs/dyn/policyanalyzer_v1.projects.locations.activityTypes.activities.html
new file mode 100644
index 00000000000..4fe17e54f20
--- /dev/null
+++ b/docs/dyn/policyanalyzer_v1.projects.locations.activityTypes.activities.html
@@ -0,0 +1,141 @@
+
+
+
+Policy Analyzer API . projects . locations . activityTypes . activities
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+ query(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+Queries policy activities on GCP resources.
+
+ query_next(previous_request, previous_response)
+Retrieves the next page of results.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ query(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+ Queries policy activities on GCP resources.
+
+Args:
+ parent: string, Required. The container resource on which to execute the request. Acceptable formats: `projects/[PROJECT_ID|PROJECT_NUMBER]/locations/[LOCATION]/activityTypes/[ACTIVITY_TYPE]` LOCATION here refers to GCP Locations: https://cloud.google.com/about/locations/ (required)
+ filter: string, Optional. Optional filter expression to restrict the activities returned. Supported filters are: - service_account_last_authn.full_resource_name {=} - service_account_key_last_authn.full_resource_name {=}
+ pageSize: integer, Optional. The maximum number of results to return from this request. Max limit is 1000. Non-positive values are ignored. The presence of `nextPageToken` in the response indicates that more results might be available.
+ pageToken: string, Optional. If present, then retrieve the next batch of results from the preceding call to this method. `pageToken` must be the value of `nextPageToken` from the previous response. The values of other method parameters should be identical to those in the previous call.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response to the `QueryActivity` method.
+ "activities": [ # The set of activities that match the filter included in the request.
+ {
+ "activity": { # A struct of custom fields to explain the activity.
+ "a_key": "", # Properties of the object.
+ },
+ "activityType": "A String", # The type of the activity.
+ "fullResourceName": "A String", # The full resource name that identifies the resource. For examples of full resource names for Google Cloud services, see https://cloud.google.com/iam/help/troubleshooter/full-resource-names.
+ "observationPeriod": { # Represents data observation period. # The data observation period to build the activity.
+ "endTime": "A String", # The observation end time.
+ "startTime": "A String", # The observation start time.
+ },
+ },
+ ],
+ "nextPageToken": "A String", # If there might be more results than those appearing in this response, then `nextPageToken` is included. To get the next set of results, call this method again using the value of `nextPageToken` as `pageToken`.
+}
+
+
+
+ query_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/policyanalyzer_v1.projects.locations.activityTypes.html b/docs/dyn/policyanalyzer_v1.projects.locations.activityTypes.html
new file mode 100644
index 00000000000..5d8421424d4
--- /dev/null
+++ b/docs/dyn/policyanalyzer_v1.projects.locations.activityTypes.html
@@ -0,0 +1,91 @@
+
+
+
+Policy Analyzer API . projects . locations . activityTypes
+Instance Methods
+
+ activities()
+
+Returns the activities Resource.
+
+
+ close()
+Close httplib2 connections.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/policyanalyzer_v1.projects.locations.html b/docs/dyn/policyanalyzer_v1.projects.locations.html
new file mode 100644
index 00000000000..3bde8b9a27f
--- /dev/null
+++ b/docs/dyn/policyanalyzer_v1.projects.locations.html
@@ -0,0 +1,91 @@
+
+
+
+Policy Analyzer API . projects . locations
+Instance Methods
+
+ activityTypes()
+
+Returns the activityTypes Resource.
+
+
+ close()
+Close httplib2 connections.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/policyanalyzer_v1beta1.html b/docs/dyn/policyanalyzer_v1beta1.html
new file mode 100644
index 00000000000..7eabc195053
--- /dev/null
+++ b/docs/dyn/policyanalyzer_v1beta1.html
@@ -0,0 +1,111 @@
+
+
+
+Policy Analyzer API
+Instance Methods
+
+ projects()
+
+Returns the projects Resource.
+
+
+ close()
+Close httplib2 connections.
+
+Create a BatchHttpRequest object based on the discovery document.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ new_batch_http_request()
+ Create a BatchHttpRequest object based on the discovery document.
+
+ Args:
+ callback: callable, A callback to be called for each response, of the
+ form callback(id, response, exception). The first parameter is the
+ request id, and the second is the deserialized response object. The
+ third is an apiclient.errors.HttpError exception object if an HTTP
+ error occurred while processing the request, or None if no error
+ occurred.
+
+ Returns:
+ A BatchHttpRequest object based on the discovery document.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/policyanalyzer_v1beta1.projects.html b/docs/dyn/policyanalyzer_v1beta1.projects.html
new file mode 100644
index 00000000000..9fbaba64cfe
--- /dev/null
+++ b/docs/dyn/policyanalyzer_v1beta1.projects.html
@@ -0,0 +1,91 @@
+
+
+
+Policy Analyzer API . projects
+Instance Methods
+
+ locations()
+
+Returns the locations Resource.
+
+
+ close()
+Close httplib2 connections.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/policyanalyzer_v1beta1.projects.locations.activityTypes.activities.html b/docs/dyn/policyanalyzer_v1beta1.projects.locations.activityTypes.activities.html
new file mode 100644
index 00000000000..383f2c56d17
--- /dev/null
+++ b/docs/dyn/policyanalyzer_v1beta1.projects.locations.activityTypes.activities.html
@@ -0,0 +1,141 @@
+
+
+
+Policy Analyzer API . projects . locations . activityTypes . activities
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+ query(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+Queries policy activities on GCP resources.
+
+ query_next(previous_request, previous_response)
+Retrieves the next page of results.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ query(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+ Queries policy activities on GCP resources.
+
+Args:
+ parent: string, Required. The container resource on which to execute the request. Acceptable formats: `projects/[PROJECT_ID|PROJECT_NUMBER]/locations/[LOCATION]/activityTypes/[ACTIVITY_TYPE]` LOCATION here refers to GCP Locations: https://cloud.google.com/about/locations/ (required)
+ filter: string, Optional. Optional filter expression to restrict the activities returned. Supported filters are: - service_account_last_authn.full_resource_name {=} - service_account_key_last_authn.full_resource_name {=}
+ pageSize: integer, Optional. The maximum number of results to return from this request. Max limit is 1000. Non-positive values are ignored. The presence of `nextPageToken` in the response indicates that more results might be available.
+ pageToken: string, Optional. If present, then retrieve the next batch of results from the preceding call to this method. `pageToken` must be the value of `nextPageToken` from the previous response. The values of other method parameters should be identical to those in the previous call.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Response to the `QueryActivity` method.
+ "activities": [ # The set of activities that match the filter included in the request.
+ {
+ "activity": { # A struct of custom fields to explain the activity.
+ "a_key": "", # Properties of the object.
+ },
+ "activityType": "A String", # The type of the activity.
+ "fullResourceName": "A String", # The full resource name that identifies the resource. For examples of full resource names for Google Cloud services, see https://cloud.google.com/iam/help/troubleshooter/full-resource-names.
+ "observationPeriod": { # Represents data observation period. # The data observation period to build the activity.
+ "endTime": "A String", # The observation end time.
+ "startTime": "A String", # The observation start time.
+ },
+ },
+ ],
+ "nextPageToken": "A String", # If there might be more results than those appearing in this response, then `nextPageToken` is included. To get the next set of results, call this method again using the value of `nextPageToken` as `pageToken`.
+}
+
+
+
+ query_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/policyanalyzer_v1beta1.projects.locations.activityTypes.html b/docs/dyn/policyanalyzer_v1beta1.projects.locations.activityTypes.html
new file mode 100644
index 00000000000..f92592474ba
--- /dev/null
+++ b/docs/dyn/policyanalyzer_v1beta1.projects.locations.activityTypes.html
@@ -0,0 +1,91 @@
+
+
+
+Policy Analyzer API . projects . locations . activityTypes
+Instance Methods
+
+ activities()
+
+Returns the activities Resource.
+
+
+ close()
+Close httplib2 connections.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/policyanalyzer_v1beta1.projects.locations.html b/docs/dyn/policyanalyzer_v1beta1.projects.locations.html
new file mode 100644
index 00000000000..ad45962c4fe
--- /dev/null
+++ b/docs/dyn/policyanalyzer_v1beta1.projects.locations.html
@@ -0,0 +1,91 @@
+
+
+
+Policy Analyzer API . projects . locations
+Instance Methods
+
+ activityTypes()
+
+Returns the activityTypes Resource.
+
+
+ close()
+Close httplib2 connections.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/privateca_v1.projects.locations.caPools.certificateAuthorities.html b/docs/dyn/privateca_v1.projects.locations.caPools.certificateAuthorities.html
index 8db42ec8e90..5b24daa160c 100644
--- a/docs/dyn/privateca_v1.projects.locations.caPools.certificateAuthorities.html
+++ b/docs/dyn/privateca_v1.projects.locations.caPools.certificateAuthorities.html
@@ -129,7 +129,7 @@ Method Details
"pemCaCertificate": "A String", # Required. The signed CA certificate issued from FetchCertificateAuthorityCsrResponse.pem_csr.
"requestId": "A String", # Optional. An ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and t he request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000).
"subordinateConfig": { # Describes a subordinate CA's issuers. This is either a resource name to a known issuing CertificateAuthority, or a PEM issuer certificate chain. # Required. Must include information about the issuer of 'pem_ca_certificate', and any further issuers until the self-signed CA.
- "certificateAuthority": "A String", # Required. This can refer to a CertificateAuthority in the same project that was used to create a subordinate CertificateAuthority. This field is used for information and usability purposes only. The resource name is in the format `projects/*/locations/*/caPools/*/certificateAuthorities/*`.
+ "certificateAuthority": "A String", # Required. This can refer to a CertificateAuthority that was used to create a subordinate CertificateAuthority. This field is used for information and usability purposes only. The resource name is in the format `projects/*/locations/*/caPools/*/certificateAuthorities/*`.
"pemIssuerChain": { # This message describes a subordinate CA's issuer certificate chain. This wrapper exists for compatibility reasons. # Required. Contains the PEM certificate chain for the issuers of this CertificateAuthority, but not pem certificate for this CA itself.
"pemCertificates": [ # Required. Expected to be in leaf-to-root order according to RFC 5246.
"A String",
@@ -423,7 +423,7 @@ Method Details
],
"state": "A String", # Output only. The State for this CertificateAuthority.
"subordinateConfig": { # Describes a subordinate CA's issuers. This is either a resource name to a known issuing CertificateAuthority, or a PEM issuer certificate chain. # Optional. If this is a subordinate CertificateAuthority, this field will be set with the subordinate configuration, which describes its issuers. This may be updated, but this CertificateAuthority must continue to validate.
- "certificateAuthority": "A String", # Required. This can refer to a CertificateAuthority in the same project that was used to create a subordinate CertificateAuthority. This field is used for information and usability purposes only. The resource name is in the format `projects/*/locations/*/caPools/*/certificateAuthorities/*`.
+ "certificateAuthority": "A String", # Required. This can refer to a CertificateAuthority that was used to create a subordinate CertificateAuthority. This field is used for information and usability purposes only. The resource name is in the format `projects/*/locations/*/caPools/*/certificateAuthorities/*`.
"pemIssuerChain": { # This message describes a subordinate CA's issuer certificate chain. This wrapper exists for compatibility reasons. # Required. Contains the PEM certificate chain for the issuers of this CertificateAuthority, but not pem certificate for this CA itself.
"pemCertificates": [ # Required. Expected to be in leaf-to-root order according to RFC 5246.
"A String",
@@ -862,7 +862,7 @@ Method Details
],
"state": "A String", # Output only. The State for this CertificateAuthority.
"subordinateConfig": { # Describes a subordinate CA's issuers. This is either a resource name to a known issuing CertificateAuthority, or a PEM issuer certificate chain. # Optional. If this is a subordinate CertificateAuthority, this field will be set with the subordinate configuration, which describes its issuers. This may be updated, but this CertificateAuthority must continue to validate.
- "certificateAuthority": "A String", # Required. This can refer to a CertificateAuthority in the same project that was used to create a subordinate CertificateAuthority. This field is used for information and usability purposes only. The resource name is in the format `projects/*/locations/*/caPools/*/certificateAuthorities/*`.
+ "certificateAuthority": "A String", # Required. This can refer to a CertificateAuthority that was used to create a subordinate CertificateAuthority. This field is used for information and usability purposes only. The resource name is in the format `projects/*/locations/*/caPools/*/certificateAuthorities/*`.
"pemIssuerChain": { # This message describes a subordinate CA's issuer certificate chain. This wrapper exists for compatibility reasons. # Required. Contains the PEM certificate chain for the issuers of this CertificateAuthority, but not pem certificate for this CA itself.
"pemCertificates": [ # Required. Expected to be in leaf-to-root order according to RFC 5246.
"A String",
@@ -1137,7 +1137,7 @@ Method Details
],
"state": "A String", # Output only. The State for this CertificateAuthority.
"subordinateConfig": { # Describes a subordinate CA's issuers. This is either a resource name to a known issuing CertificateAuthority, or a PEM issuer certificate chain. # Optional. If this is a subordinate CertificateAuthority, this field will be set with the subordinate configuration, which describes its issuers. This may be updated, but this CertificateAuthority must continue to validate.
- "certificateAuthority": "A String", # Required. This can refer to a CertificateAuthority in the same project that was used to create a subordinate CertificateAuthority. This field is used for information and usability purposes only. The resource name is in the format `projects/*/locations/*/caPools/*/certificateAuthorities/*`.
+ "certificateAuthority": "A String", # Required. This can refer to a CertificateAuthority that was used to create a subordinate CertificateAuthority. This field is used for information and usability purposes only. The resource name is in the format `projects/*/locations/*/caPools/*/certificateAuthorities/*`.
"pemIssuerChain": { # This message describes a subordinate CA's issuer certificate chain. This wrapper exists for compatibility reasons. # Required. Contains the PEM certificate chain for the issuers of this CertificateAuthority, but not pem certificate for this CA itself.
"pemCertificates": [ # Required. Expected to be in leaf-to-root order according to RFC 5246.
"A String",
@@ -1421,7 +1421,7 @@ Method Details
],
"state": "A String", # Output only. The State for this CertificateAuthority.
"subordinateConfig": { # Describes a subordinate CA's issuers. This is either a resource name to a known issuing CertificateAuthority, or a PEM issuer certificate chain. # Optional. If this is a subordinate CertificateAuthority, this field will be set with the subordinate configuration, which describes its issuers. This may be updated, but this CertificateAuthority must continue to validate.
- "certificateAuthority": "A String", # Required. This can refer to a CertificateAuthority in the same project that was used to create a subordinate CertificateAuthority. This field is used for information and usability purposes only. The resource name is in the format `projects/*/locations/*/caPools/*/certificateAuthorities/*`.
+ "certificateAuthority": "A String", # Required. This can refer to a CertificateAuthority that was used to create a subordinate CertificateAuthority. This field is used for information and usability purposes only. The resource name is in the format `projects/*/locations/*/caPools/*/certificateAuthorities/*`.
"pemIssuerChain": { # This message describes a subordinate CA's issuer certificate chain. This wrapper exists for compatibility reasons. # Required. Contains the PEM certificate chain for the issuers of this CertificateAuthority, but not pem certificate for this CA itself.
"pemCertificates": [ # Required. Expected to be in leaf-to-root order according to RFC 5246.
"A String",
diff --git a/docs/dyn/pubsub_v1.projects.subscriptions.html b/docs/dyn/pubsub_v1.projects.subscriptions.html
index 70ddb0152b8..bb524326723 100644
--- a/docs/dyn/pubsub_v1.projects.subscriptions.html
+++ b/docs/dyn/pubsub_v1.projects.subscriptions.html
@@ -197,6 +197,7 @@ Method Details
"minimumBackoff": "A String", # The minimum delay between consecutive deliveries of a given message. Value should be between 0 and 600 seconds. Defaults to 10 seconds.
},
"topic": "A String", # Required. The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. The value of this field will be `_deleted-topic_` if the topic has been deleted.
+ "topicMessageRetentionDuration": "A String", # Output only. Indicates the minimum duration for which a message is retained after it is published to the subscription's topic. If this field is set, messages published to the subscription's topic in the last `topic_message_retention_duration` are always available to subscribers. See the `message_retention_duration` field in `Topic`. This field is set only in responses from the server; it is ignored if it is set in any requests.
}
x__xgafv: string, V1 error format.
@@ -240,6 +241,7 @@ Method Details
"minimumBackoff": "A String", # The minimum delay between consecutive deliveries of a given message. Value should be between 0 and 600 seconds. Defaults to 10 seconds.
},
"topic": "A String", # Required. The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. The value of this field will be `_deleted-topic_` if the topic has been deleted.
+ "topicMessageRetentionDuration": "A String", # Output only. Indicates the minimum duration for which a message is retained after it is published to the subscription's topic. If this field is set, messages published to the subscription's topic in the last `topic_message_retention_duration` are always available to subscribers. See the `message_retention_duration` field in `Topic`. This field is set only in responses from the server; it is ignored if it is set in any requests.
}
@@ -326,6 +328,7 @@ Method Details
"minimumBackoff": "A String", # The minimum delay between consecutive deliveries of a given message. Value should be between 0 and 600 seconds. Defaults to 10 seconds.
},
"topic": "A String", # Required. The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. The value of this field will be `_deleted-topic_` if the topic has been deleted.
+ "topicMessageRetentionDuration": "A String", # Output only. Indicates the minimum duration for which a message is retained after it is published to the subscription's topic. If this field is set, messages published to the subscription's topic in the last `topic_message_retention_duration` are always available to subscribers. See the `message_retention_duration` field in `Topic`. This field is set only in responses from the server; it is ignored if it is set in any requests.
}
@@ -416,6 +419,7 @@ Method Details
"minimumBackoff": "A String", # The minimum delay between consecutive deliveries of a given message. Value should be between 0 and 600 seconds. Defaults to 10 seconds.
},
"topic": "A String", # Required. The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. The value of this field will be `_deleted-topic_` if the topic has been deleted.
+ "topicMessageRetentionDuration": "A String", # Output only. Indicates the minimum duration for which a message is retained after it is published to the subscription's topic. If this field is set, messages published to the subscription's topic in the last `topic_message_retention_duration` are always available to subscribers. See the `message_retention_duration` field in `Topic`. This field is set only in responses from the server; it is ignored if it is set in any requests.
},
],
}
@@ -540,6 +544,7 @@ Method Details
"minimumBackoff": "A String", # The minimum delay between consecutive deliveries of a given message. Value should be between 0 and 600 seconds. Defaults to 10 seconds.
},
"topic": "A String", # Required. The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. The value of this field will be `_deleted-topic_` if the topic has been deleted.
+ "topicMessageRetentionDuration": "A String", # Output only. Indicates the minimum duration for which a message is retained after it is published to the subscription's topic. If this field is set, messages published to the subscription's topic in the last `topic_message_retention_duration` are always available to subscribers. See the `message_retention_duration` field in `Topic`. This field is set only in responses from the server; it is ignored if it is set in any requests.
},
"updateMask": "A String", # Required. Indicates which fields in the provided subscription to update. Must be specified and non-empty.
}
@@ -585,6 +590,7 @@ Method Details
"minimumBackoff": "A String", # The minimum delay between consecutive deliveries of a given message. Value should be between 0 and 600 seconds. Defaults to 10 seconds.
},
"topic": "A String", # Required. The name of the topic from which this subscription is receiving messages. Format is `projects/{project}/topics/{topic}`. The value of this field will be `_deleted-topic_` if the topic has been deleted.
+ "topicMessageRetentionDuration": "A String", # Output only. Indicates the minimum duration for which a message is retained after it is published to the subscription's topic. If this field is set, messages published to the subscription's topic in the last `topic_message_retention_duration` are always available to subscribers. See the `message_retention_duration` field in `Topic`. This field is set only in responses from the server; it is ignored if it is set in any requests.
}
diff --git a/docs/dyn/pubsub_v1.projects.topics.html b/docs/dyn/pubsub_v1.projects.topics.html
index c7f51b1cf93..30ae568adbd 100644
--- a/docs/dyn/pubsub_v1.projects.topics.html
+++ b/docs/dyn/pubsub_v1.projects.topics.html
@@ -137,6 +137,7 @@ Method Details
"labels": { # See [Creating and managing labels] (https://cloud.google.com/pubsub/docs/labels).
"a_key": "A String",
},
+ "messageRetentionDuration": "A String", # Indicates the minimum duration to retain a message after it is published to the topic. If this field is set, messages published to the topic in the last `message_retention_duration` are always available to subscribers. For instance, it allows any attached subscription to [seek to a timestamp](https://cloud.google.com/pubsub/docs/replay-overview#seek_to_a_time) that is up to `message_retention_duration` in the past. If this field is not set, message retention is controlled by settings on individual subscriptions. Cannot be more than 7 days or less than 10 minutes.
"messageStoragePolicy": { # A policy constraining the storage of messages published to the topic. # Policy constraining the set of Google Cloud Platform regions where messages published to the topic may be stored. If not present, then no constraints are in effect.
"allowedPersistenceRegions": [ # A list of IDs of GCP regions where messages that are published to the topic may be persisted in storage. Messages published by publishers running in non-allowed GCP regions (or running outside of GCP altogether) will be routed for storage in one of the allowed regions. An empty list means that no regions are allowed, and is not a valid configuration.
"A String",
@@ -163,6 +164,7 @@ Method Details
"labels": { # See [Creating and managing labels] (https://cloud.google.com/pubsub/docs/labels).
"a_key": "A String",
},
+ "messageRetentionDuration": "A String", # Indicates the minimum duration to retain a message after it is published to the topic. If this field is set, messages published to the topic in the last `message_retention_duration` are always available to subscribers. For instance, it allows any attached subscription to [seek to a timestamp](https://cloud.google.com/pubsub/docs/replay-overview#seek_to_a_time) that is up to `message_retention_duration` in the past. If this field is not set, message retention is controlled by settings on individual subscriptions. Cannot be more than 7 days or less than 10 minutes.
"messageStoragePolicy": { # A policy constraining the storage of messages published to the topic. # Policy constraining the set of Google Cloud Platform regions where messages published to the topic may be stored. If not present, then no constraints are in effect.
"allowedPersistenceRegions": [ # A list of IDs of GCP regions where messages that are published to the topic may be persisted in storage. Messages published by publishers running in non-allowed GCP regions (or running outside of GCP altogether) will be routed for storage in one of the allowed regions. An empty list means that no regions are allowed, and is not a valid configuration.
"A String",
@@ -214,6 +216,7 @@ Method Details
"labels": { # See [Creating and managing labels] (https://cloud.google.com/pubsub/docs/labels).
"a_key": "A String",
},
+ "messageRetentionDuration": "A String", # Indicates the minimum duration to retain a message after it is published to the topic. If this field is set, messages published to the topic in the last `message_retention_duration` are always available to subscribers. For instance, it allows any attached subscription to [seek to a timestamp](https://cloud.google.com/pubsub/docs/replay-overview#seek_to_a_time) that is up to `message_retention_duration` in the past. If this field is not set, message retention is controlled by settings on individual subscriptions. Cannot be more than 7 days or less than 10 minutes.
"messageStoragePolicy": { # A policy constraining the storage of messages published to the topic. # Policy constraining the set of Google Cloud Platform regions where messages published to the topic may be stored. If not present, then no constraints are in effect.
"allowedPersistenceRegions": [ # A list of IDs of GCP regions where messages that are published to the topic may be persisted in storage. Messages published by publishers running in non-allowed GCP regions (or running outside of GCP altogether) will be routed for storage in one of the allowed regions. An empty list means that no regions are allowed, and is not a valid configuration.
"A String",
@@ -287,6 +290,7 @@ Method Details
"labels": { # See [Creating and managing labels] (https://cloud.google.com/pubsub/docs/labels).
"a_key": "A String",
},
+ "messageRetentionDuration": "A String", # Indicates the minimum duration to retain a message after it is published to the topic. If this field is set, messages published to the topic in the last `message_retention_duration` are always available to subscribers. For instance, it allows any attached subscription to [seek to a timestamp](https://cloud.google.com/pubsub/docs/replay-overview#seek_to_a_time) that is up to `message_retention_duration` in the past. If this field is not set, message retention is controlled by settings on individual subscriptions. Cannot be more than 7 days or less than 10 minutes.
"messageStoragePolicy": { # A policy constraining the storage of messages published to the topic. # Policy constraining the set of Google Cloud Platform regions where messages published to the topic may be stored. If not present, then no constraints are in effect.
"allowedPersistenceRegions": [ # A list of IDs of GCP regions where messages that are published to the topic may be persisted in storage. Messages published by publishers running in non-allowed GCP regions (or running outside of GCP altogether) will be routed for storage in one of the allowed regions. An empty list means that no regions are allowed, and is not a valid configuration.
"A String",
@@ -332,6 +336,7 @@ Method Details
"labels": { # See [Creating and managing labels] (https://cloud.google.com/pubsub/docs/labels).
"a_key": "A String",
},
+ "messageRetentionDuration": "A String", # Indicates the minimum duration to retain a message after it is published to the topic. If this field is set, messages published to the topic in the last `message_retention_duration` are always available to subscribers. For instance, it allows any attached subscription to [seek to a timestamp](https://cloud.google.com/pubsub/docs/replay-overview#seek_to_a_time) that is up to `message_retention_duration` in the past. If this field is not set, message retention is controlled by settings on individual subscriptions. Cannot be more than 7 days or less than 10 minutes.
"messageStoragePolicy": { # A policy constraining the storage of messages published to the topic. # Policy constraining the set of Google Cloud Platform regions where messages published to the topic may be stored. If not present, then no constraints are in effect.
"allowedPersistenceRegions": [ # A list of IDs of GCP regions where messages that are published to the topic may be persisted in storage. Messages published by publishers running in non-allowed GCP regions (or running outside of GCP altogether) will be routed for storage in one of the allowed regions. An empty list means that no regions are allowed, and is not a valid configuration.
"A String",
@@ -360,6 +365,7 @@ Method Details
"labels": { # See [Creating and managing labels] (https://cloud.google.com/pubsub/docs/labels).
"a_key": "A String",
},
+ "messageRetentionDuration": "A String", # Indicates the minimum duration to retain a message after it is published to the topic. If this field is set, messages published to the topic in the last `message_retention_duration` are always available to subscribers. For instance, it allows any attached subscription to [seek to a timestamp](https://cloud.google.com/pubsub/docs/replay-overview#seek_to_a_time) that is up to `message_retention_duration` in the past. If this field is not set, message retention is controlled by settings on individual subscriptions. Cannot be more than 7 days or less than 10 minutes.
"messageStoragePolicy": { # A policy constraining the storage of messages published to the topic. # Policy constraining the set of Google Cloud Platform regions where messages published to the topic may be stored. If not present, then no constraints are in effect.
"allowedPersistenceRegions": [ # A list of IDs of GCP regions where messages that are published to the topic may be persisted in storage. Messages published by publishers running in non-allowed GCP regions (or running outside of GCP altogether) will be routed for storage in one of the allowed regions. An empty list means that no regions are allowed, and is not a valid configuration.
"A String",
diff --git a/docs/dyn/pubsublite_v1.admin.projects.locations.html b/docs/dyn/pubsublite_v1.admin.projects.locations.html
index 69a117e4dfe..9cb61946f27 100644
--- a/docs/dyn/pubsublite_v1.admin.projects.locations.html
+++ b/docs/dyn/pubsublite_v1.admin.projects.locations.html
@@ -74,6 +74,11 @@
Pub/Sub Lite API . admin . projects . locations
Instance Methods
+
+ operations()
+
+Returns the operations Resource.
+
diff --git a/docs/dyn/pubsublite_v1.admin.projects.locations.operations.html b/docs/dyn/pubsublite_v1.admin.projects.locations.operations.html
new file mode 100644
index 00000000000..02b1c344fae
--- /dev/null
+++ b/docs/dyn/pubsublite_v1.admin.projects.locations.operations.html
@@ -0,0 +1,235 @@
+
+
+
+Pub/Sub Lite API . admin . projects . locations . operations
+Instance Methods
+
+ cancel(name, body=None, x__xgafv=None)
+Starts asynchronous cancellation on a long-running operation. The server makes a best effort to cancel the operation, but success is not guaranteed. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`. Clients can use Operations.GetOperation or other methods to check whether the cancellation succeeded or whether the operation completed despite cancellation. On successful cancellation, the operation is not deleted; instead, it becomes an operation with an Operation.error value with a google.rpc.Status.code of 1, corresponding to `Code.CANCELLED`.
+
+ close()
+Close httplib2 connections.
+
+Deletes a long-running operation. This method indicates that the client is no longer interested in the operation result. It does not cancel the operation. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`.
+
+Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+ list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `"/v1/{name=users/*}/operations"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.
+
+ list_next(previous_request, previous_response)
+Retrieves the next page of results.
+Method Details
+
+ cancel(name, body=None, x__xgafv=None)
+ Starts asynchronous cancellation on a long-running operation. The server makes a best effort to cancel the operation, but success is not guaranteed. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`. Clients can use Operations.GetOperation or other methods to check whether the cancellation succeeded or whether the operation completed despite cancellation. On successful cancellation, the operation is not deleted; instead, it becomes an operation with an Operation.error value with a google.rpc.Status.code of 1, corresponding to `Code.CANCELLED`.
+
+Args:
+ name: string, The name of the operation resource to be cancelled. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # The request message for Operations.CancelOperation.
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`.
+}
+
+
+
+ close()
+ Close httplib2 connections.
+
+
+
+ delete(name, x__xgafv=None)
+ Deletes a long-running operation. This method indicates that the client is no longer interested in the operation result. It does not cancel the operation. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`.
+
+Args:
+ name: string, The name of the operation resource to be deleted. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`.
+}
+
+
+
+ get(name, x__xgafv=None)
+ Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
+
+Args:
+ name: string, The name of the operation resource. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+
+
+
+ list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+ Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `"/v1/{name=users/*}/operations"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.
+
+Args:
+ name: string, The name of the operation's parent resource. (required)
+ filter: string, The standard list filter.
+ pageSize: integer, The standard list page size.
+ pageToken: string, The standard list page token.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # The response message for Operations.ListOperations.
+ "nextPageToken": "A String", # The standard List next-page token.
+ "operations": [ # A list of operations that matches the specified filter in the request.
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ },
+ ],
+}
+
+
+
+ list_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/pubsublite_v1.admin.projects.locations.subscriptions.html b/docs/dyn/pubsublite_v1.admin.projects.locations.subscriptions.html
index 480b409eefe..ae0deb0f5df 100644
--- a/docs/dyn/pubsublite_v1.admin.projects.locations.subscriptions.html
+++ b/docs/dyn/pubsublite_v1.admin.projects.locations.subscriptions.html
@@ -95,6 +95,9 @@ Instance Methods
patch(name, body=None, updateMask=None, x__xgafv=None)
Updates properties of the specified subscription.
+
+ seek(name, body=None, x__xgafv=None)
+Performs an out-of-band seek for a subscription to a specified target, which may be timestamps or named positions within the message backlog. Seek translates these targets to cursors for each partition and orchestrates subscribers to start consuming messages from these seek cursors. If an operation is returned, the seek has been registered and subscribers will eventually receive messages from the seek cursors (i.e. eventual consistency), as long as they are using a minimum supported client library version and not a system that tracks cursors independently of Pub/Sub Lite (e.g. Apache Beam, Dataflow, Spark). The seek operation will fail for unsupported clients. If clients would like to know when subscribers react to the seek (or not), they can poll the operation. The seek operation will succeed and complete once subscribers are ready to receive messages from the seek cursors for all partitions of the topic. This means that the seek operation will not complete until all subscribers come online. If the previous seek operation has not yet completed, it will be aborted and the new invocation of seek will supersede it.
Method Details
close()
@@ -257,4 +260,50 @@ Method Details
}
+
+ seek(name, body=None, x__xgafv=None)
+ Performs an out-of-band seek for a subscription to a specified target, which may be timestamps or named positions within the message backlog. Seek translates these targets to cursors for each partition and orchestrates subscribers to start consuming messages from these seek cursors. If an operation is returned, the seek has been registered and subscribers will eventually receive messages from the seek cursors (i.e. eventual consistency), as long as they are using a minimum supported client library version and not a system that tracks cursors independently of Pub/Sub Lite (e.g. Apache Beam, Dataflow, Spark). The seek operation will fail for unsupported clients. If clients would like to know when subscribers react to the seek (or not), they can poll the operation. The seek operation will succeed and complete once subscribers are ready to receive messages from the seek cursors for all partitions of the topic. This means that the seek operation will not complete until all subscribers come online. If the previous seek operation has not yet completed, it will be aborted and the new invocation of seek will supersede it.
+
+Args:
+ name: string, Required. The name of the subscription to seek. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request for SeekSubscription.
+ "namedTarget": "A String", # Seek to a named position with respect to the message backlog.
+ "timeTarget": { # A target publish or event time. Can be used for seeking to or retrieving the corresponding cursor. # Seek to the first message whose publish or event time is greater than or equal to the specified query time. If no such message can be located, will seek to the end of the message backlog.
+ "eventTime": "A String", # Request the cursor of the first message with event time greater than or equal to `event_time`. If messages are missing an event time, the publish time is used as a fallback. As event times are user supplied, subsequent messages may have event times less than `event_time` and should be filtered by the client, if necessary.
+ "publishTime": "A String", # Request the cursor of the first message with publish time greater than or equal to `publish_time`. All messages thereafter are guaranteed to have publish times >= `publish_time`.
+ },
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+
+