diff --git a/.flake8 b/.flake8 index 29227d4c..2e438749 100644 --- a/.flake8 +++ b/.flake8 @@ -16,7 +16,7 @@ # Generated by synthtool. DO NOT EDIT! [flake8] -ignore = E203, E266, E501, W503 +ignore = E203, E231, E266, E501, W503 exclude = # Exclude generated code. **/proto/** diff --git a/.github/.OwlBot.lock.yaml b/.github/.OwlBot.lock.yaml index 44c78f7c..757c9dca 100644 --- a/.github/.OwlBot.lock.yaml +++ b/.github/.OwlBot.lock.yaml @@ -13,4 +13,5 @@ # limitations under the License. docker: image: gcr.io/cloud-devrel-public-resources/owlbot-python:latest - digest: sha256:4e1991042fe54b991db9ca17c8fb386e61b22fe4d1472a568bf0fcac85dcf5d3 + digest: sha256:81ed5ecdfc7cac5b699ba4537376f3563f6f04122c4ec9e735d3b3dc1d43dd32 +# created: 2022-05-05T22:08:23.383410683Z diff --git a/.github/auto-approve.yml b/.github/auto-approve.yml new file mode 100644 index 00000000..311ebbb8 --- /dev/null +++ b/.github/auto-approve.yml @@ -0,0 +1,3 @@ +# https://github.com/googleapis/repo-automation-bots/tree/main/packages/auto-approve +processes: + - "OwlBotTemplateChanges" diff --git a/.github/auto-label.yaml b/.github/auto-label.yaml new file mode 100644 index 00000000..41bff0b5 --- /dev/null +++ b/.github/auto-label.yaml @@ -0,0 +1,15 @@ +# Copyright 2022 Google LLC +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +requestsize: + enabled: true diff --git a/.github/release-please.yml b/.github/release-please.yml index 466597e5..29601ad4 100644 --- a/.github/release-please.yml +++ b/.github/release-please.yml @@ -1,2 +1,11 @@ releaseType: python handleGHRelease: true +# NOTE: this section is generated by synthtool.languages.python +# See https://github.com/googleapis/synthtool/blob/master/synthtool/languages/python.py +branches: +- branch: v1 + handleGHRelease: true + releaseType: python +- branch: v0 + handleGHRelease: true + releaseType: python diff --git a/.kokoro/docker/docs/Dockerfile b/.kokoro/docker/docs/Dockerfile index 4e1b1fb8..238b87b9 100644 --- a/.kokoro/docker/docs/Dockerfile +++ b/.kokoro/docker/docs/Dockerfile @@ -12,7 +12,7 @@ # See the License for the specific language governing permissions and # limitations under the License. -from ubuntu:20.04 +from ubuntu:22.04 ENV DEBIAN_FRONTEND noninteractive @@ -60,8 +60,24 @@ RUN apt-get update \ && rm -rf /var/lib/apt/lists/* \ && rm -f /var/cache/apt/archives/*.deb +###################### Install python 3.8.11 + +# Download python 3.8.11 +RUN wget https://www.python.org/ftp/python/3.8.11/Python-3.8.11.tgz + +# Extract files +RUN tar -xvf Python-3.8.11.tgz + +# Install python 3.8.11 +RUN ./Python-3.8.11/configure --enable-optimizations +RUN make altinstall + +###################### Install pip RUN wget -O /tmp/get-pip.py 'https://bootstrap.pypa.io/get-pip.py' \ - && python3.8 /tmp/get-pip.py \ + && python3 /tmp/get-pip.py \ && rm /tmp/get-pip.py +# Test pip +RUN python3 -m pip + CMD ["python3.8"] diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index 62eb5a77..46d23716 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -22,7 +22,7 @@ repos: - id: end-of-file-fixer - id: check-yaml - repo: https://github.com/psf/black - rev: 19.10b0 + rev: 22.3.0 hooks: - id: black - repo: https://gitlab.com/pycqa/flake8 diff --git a/CHANGELOG.md b/CHANGELOG.md index 4e359255..78dd07cd 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,14 +4,26 @@ [1]: https://pypi.org/project/google-cloud-container/#history -### [2.10.7](https://github.com/googleapis/python-container/compare/v2.10.6...v2.10.7) (2022-03-22) +## [2.10.8](https://github.com/googleapis/python-container/compare/v2.10.7...v2.10.8) (2022-06-03) + + +### Bug Fixes + +* **deps:** require protobuf <4.0.0dev ([#252](https://github.com/googleapis/python-container/issues/252)) ([f5ab2a8](https://github.com/googleapis/python-container/commit/f5ab2a89d1238b2315963f4dd8469323746459f0)) + + +### Documentation + +* fix changelog header to consistent size ([#253](https://github.com/googleapis/python-container/issues/253)) ([9db4c78](https://github.com/googleapis/python-container/commit/9db4c786430da9a5831893a31321cb0e65db4751)) + +## [2.10.7](https://github.com/googleapis/python-container/compare/v2.10.6...v2.10.7) (2022-03-22) ### Bug Fixes * test cleanup stages with try finally ([#212](https://github.com/googleapis/python-container/issues/212)) ([529bcbf](https://github.com/googleapis/python-container/commit/529bcbf618858aab17b6f5e86d25069a1266860a)) -### [2.10.6](https://github.com/googleapis/python-container/compare/v2.10.5...v2.10.6) (2022-03-07) +## [2.10.6](https://github.com/googleapis/python-container/compare/v2.10.5...v2.10.6) (2022-03-07) ### Bug Fixes @@ -19,34 +31,34 @@ * **deps:** require google-api-core>=1.31.5, >=2.3.2 ([#202](https://github.com/googleapis/python-container/issues/202)) ([444b806](https://github.com/googleapis/python-container/commit/444b8065a22da4c261b8b33ae8564d8329d3435d)) * **deps:** require proto-plus>=1.15.0 ([444b806](https://github.com/googleapis/python-container/commit/444b8065a22da4c261b8b33ae8564d8329d3435d)) -### [2.10.5](https://github.com/googleapis/python-container/compare/v2.10.4...v2.10.5) (2022-02-16) +## [2.10.5](https://github.com/googleapis/python-container/compare/v2.10.4...v2.10.5) (2022-02-16) ### Documentation * **samples:** add usage samples to show handling of LRO response Operation ([#191](https://github.com/googleapis/python-container/issues/191)) ([309ad62](https://github.com/googleapis/python-container/commit/309ad6219a6e80d08bcd365a163e8273a6413ede)) -### [2.10.4](https://github.com/googleapis/python-container/compare/v2.10.3...v2.10.4) (2022-02-14) +## [2.10.4](https://github.com/googleapis/python-container/compare/v2.10.3...v2.10.4) (2022-02-14) ### Bug Fixes * **deps:** move libcst to extras ([#194](https://github.com/googleapis/python-container/issues/194)) ([1c308c2](https://github.com/googleapis/python-container/commit/1c308c2e44dc16d0e8df5976de0b65d1e7c2041e)) -### [2.10.3](https://github.com/googleapis/python-container/compare/v2.10.2...v2.10.3) (2022-02-11) +## [2.10.3](https://github.com/googleapis/python-container/compare/v2.10.2...v2.10.3) (2022-02-11) ### Documentation * add generated snippets ([#192](https://github.com/googleapis/python-container/issues/192)) ([e3a3a05](https://github.com/googleapis/python-container/commit/e3a3a056d80ac713edbf5cb4a8358063f8a83214)) -### [2.10.2](https://github.com/googleapis/python-container/compare/v2.10.1...v2.10.2) (2022-02-04) +## [2.10.2](https://github.com/googleapis/python-container/compare/v2.10.1...v2.10.2) (2022-02-04) ### Bug Fixes * resolve DuplicateCredentialArgs error when using credentials_file ([07fbf3c](https://github.com/googleapis/python-container/commit/07fbf3cb1e140abf020e7cfbd083ed79aae701bf)) -### [2.10.1](https://www.github.com/googleapis/python-container/compare/v2.10.0...v2.10.1) (2021-11-01) +## [2.10.1](https://www.github.com/googleapis/python-container/compare/v2.10.0...v2.10.1) (2021-11-01) ### Bug Fixes @@ -73,7 +85,7 @@ * add context manager support in client ([#157](https://www.github.com/googleapis/python-container/issues/157)) ([03953f8](https://www.github.com/googleapis/python-container/commit/03953f8087b2583369b877672be81f2b8638020c)) -### [2.8.1](https://www.github.com/googleapis/python-container/compare/v2.8.0...v2.8.1) (2021-10-04) +## [2.8.1](https://www.github.com/googleapis/python-container/compare/v2.8.0...v2.8.1) (2021-10-04) ### Bug Fixes @@ -106,7 +118,7 @@ * clarified SetNodePoolSize API behavior ([616b21a](https://www.github.com/googleapis/python-container/commit/616b21a6abe2b0c4dd647cf56d544c2aff7312f7)) -### [2.7.1](https://www.github.com/googleapis/python-container/compare/v2.7.0...v2.7.1) (2021-07-24) +## [2.7.1](https://www.github.com/googleapis/python-container/compare/v2.7.0...v2.7.1) (2021-07-24) ### Bug Fixes @@ -120,7 +132,7 @@ * add Samples section to CONTRIBUTING.rst ([#129](https://www.github.com/googleapis/python-container/issues/129)) ([a5905b8](https://www.github.com/googleapis/python-container/commit/a5905b820c970217a3ad1604982a7e38412d8dda)) -### [2.6.1](https://www.github.com/googleapis/python-container/compare/v2.6.0...v2.6.1) (2021-07-20) +## [2.6.1](https://www.github.com/googleapis/python-container/compare/v2.6.0...v2.6.1) (2021-07-20) ### Bug Fixes @@ -148,7 +160,7 @@ * omit mention of Python 2.7 in 'CONTRIBUTING.rst' ([#1127](https://www.github.com/googleapis/python-container/issues/1127)) ([#112](https://www.github.com/googleapis/python-container/issues/112)) ([5a3941a](https://www.github.com/googleapis/python-container/commit/5a3941a03c48e3cef4d21ac10fc8e7b1b594ad1e)), closes [#1126](https://www.github.com/googleapis/python-container/issues/1126) -### [2.4.1](https://www.github.com/googleapis/python-container/compare/v2.4.0...v2.4.1) (2021-05-16) +## [2.4.1](https://www.github.com/googleapis/python-container/compare/v2.4.0...v2.4.1) (2021-05-16) ### Bug Fixes @@ -162,7 +174,7 @@ * add `from_service_account_info` ([#66](https://www.github.com/googleapis/python-container/issues/66)) ([f4f154d](https://www.github.com/googleapis/python-container/commit/f4f154db737fed68c49303133f4479199c0fdb12)) -### [2.3.1](https://www.github.com/googleapis/python-container/compare/v2.3.0...v2.3.1) (2021-02-18) +## [2.3.1](https://www.github.com/googleapis/python-container/compare/v2.3.0...v2.3.1) (2021-02-18) ### Bug Fixes @@ -226,7 +238,7 @@ All changes are from [#51](https://www.github.com/googleapis/python-container/is * regenerate client lib to pick up new mtls env ([#44](https://www.github.com/googleapis/python-container/issues/44)) ([c4ffea0](https://www.github.com/googleapis/python-container/commit/c4ffea02fbc6c6566a4e772e2b353a5b4dc5b2fc)) -### [2.0.1](https://www.github.com/googleapis/python-container/compare/v2.0.0...v2.0.1) (2020-07-24) +## [2.0.1](https://www.github.com/googleapis/python-container/compare/v2.0.0...v2.0.1) (2020-07-24) ### Bug Fixes @@ -249,7 +261,7 @@ All changes are from [#51](https://www.github.com/googleapis/python-container/is * migrate to microgenerator ([#33](https://www.github.com/googleapis/python-container/issues/33)) ([aa9b20c](https://www.github.com/googleapis/python-container/commit/aa9b20c6f4ccb6dff305bfcd72e1bde4a1ee86cd)) -### [1.0.1](https://www.github.com/googleapis/python-container/compare/v1.0.0...v1.0.1) (2020-06-16) +## [1.0.1](https://www.github.com/googleapis/python-container/compare/v1.0.0...v1.0.1) (2020-06-16) ### Bug Fixes diff --git a/docs/conf.py b/docs/conf.py index 39425405..b72c85ec 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -24,9 +24,9 @@ # All configuration values have a default; values that are commented out # serve to show the default. -import sys import os import shlex +import sys # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the @@ -361,7 +361,10 @@ intersphinx_mapping = { "python": ("https://python.readthedocs.org/en/latest/", None), "google-auth": ("https://googleapis.dev/python/google-auth/latest/", None), - "google.api_core": ("https://googleapis.dev/python/google-api-core/latest/", None,), + "google.api_core": ( + "https://googleapis.dev/python/google-api-core/latest/", + None, + ), "grpc": ("https://grpc.github.io/grpc/python/", None), "proto-plus": ("https://proto-plus-python.readthedocs.io/en/latest/", None), "protobuf": ("https://googleapis.dev/python/protobuf/latest/", None), diff --git a/google/cloud/container/__init__.py b/google/cloud/container/__init__.py index a84e021a..59467da8 100644 --- a/google/cloud/container/__init__.py +++ b/google/cloud/container/__init__.py @@ -14,152 +14,137 @@ # limitations under the License. # -from google.cloud.container_v1.services.cluster_manager.client import ( - ClusterManagerClient, -) from google.cloud.container_v1.services.cluster_manager.async_client import ( ClusterManagerAsyncClient, ) - -from google.cloud.container_v1.types.cluster_service import AcceleratorConfig -from google.cloud.container_v1.types.cluster_service import AddonsConfig -from google.cloud.container_v1.types.cluster_service import AdvancedMachineFeatures -from google.cloud.container_v1.types.cluster_service import AuthenticatorGroupsConfig -from google.cloud.container_v1.types.cluster_service import Autopilot -from google.cloud.container_v1.types.cluster_service import ( - AutoprovisioningNodePoolDefaults, +from google.cloud.container_v1.services.cluster_manager.client import ( + ClusterManagerClient, ) -from google.cloud.container_v1.types.cluster_service import AutoUpgradeOptions -from google.cloud.container_v1.types.cluster_service import BinaryAuthorization -from google.cloud.container_v1.types.cluster_service import CancelOperationRequest -from google.cloud.container_v1.types.cluster_service import ClientCertificateConfig -from google.cloud.container_v1.types.cluster_service import CloudRunConfig -from google.cloud.container_v1.types.cluster_service import Cluster -from google.cloud.container_v1.types.cluster_service import ClusterAutoscaling -from google.cloud.container_v1.types.cluster_service import ClusterUpdate -from google.cloud.container_v1.types.cluster_service import CompleteIPRotationRequest -from google.cloud.container_v1.types.cluster_service import ConfidentialNodes -from google.cloud.container_v1.types.cluster_service import ConfigConnectorConfig -from google.cloud.container_v1.types.cluster_service import CreateClusterRequest -from google.cloud.container_v1.types.cluster_service import CreateNodePoolRequest -from google.cloud.container_v1.types.cluster_service import DailyMaintenanceWindow -from google.cloud.container_v1.types.cluster_service import DatabaseEncryption -from google.cloud.container_v1.types.cluster_service import DefaultSnatStatus -from google.cloud.container_v1.types.cluster_service import DeleteClusterRequest -from google.cloud.container_v1.types.cluster_service import DeleteNodePoolRequest -from google.cloud.container_v1.types.cluster_service import DnsCacheConfig -from google.cloud.container_v1.types.cluster_service import DNSConfig from google.cloud.container_v1.types.cluster_service import ( + AcceleratorConfig, + AddonsConfig, + AdvancedMachineFeatures, + AuthenticatorGroupsConfig, + Autopilot, + AutoprovisioningNodePoolDefaults, + AutoUpgradeOptions, + BinaryAuthorization, + CancelOperationRequest, + ClientCertificateConfig, + CloudRunConfig, + Cluster, + ClusterAutoscaling, + ClusterUpdate, + CompleteIPRotationRequest, + ConfidentialNodes, + ConfigConnectorConfig, + CreateClusterRequest, + CreateNodePoolRequest, + DailyMaintenanceWindow, + DatabaseEncryption, + DatapathProvider, + DefaultSnatStatus, + DeleteClusterRequest, + DeleteNodePoolRequest, + DnsCacheConfig, + DNSConfig, GcePersistentDiskCsiDriverConfig, -) -from google.cloud.container_v1.types.cluster_service import GcfsConfig -from google.cloud.container_v1.types.cluster_service import GcpFilestoreCsiDriverConfig -from google.cloud.container_v1.types.cluster_service import GetClusterRequest -from google.cloud.container_v1.types.cluster_service import GetJSONWebKeysRequest -from google.cloud.container_v1.types.cluster_service import GetJSONWebKeysResponse -from google.cloud.container_v1.types.cluster_service import GetNodePoolRequest -from google.cloud.container_v1.types.cluster_service import GetOpenIDConfigRequest -from google.cloud.container_v1.types.cluster_service import GetOpenIDConfigResponse -from google.cloud.container_v1.types.cluster_service import GetOperationRequest -from google.cloud.container_v1.types.cluster_service import GetServerConfigRequest -from google.cloud.container_v1.types.cluster_service import HorizontalPodAutoscaling -from google.cloud.container_v1.types.cluster_service import HttpLoadBalancing -from google.cloud.container_v1.types.cluster_service import ILBSubsettingConfig -from google.cloud.container_v1.types.cluster_service import IntraNodeVisibilityConfig -from google.cloud.container_v1.types.cluster_service import IPAllocationPolicy -from google.cloud.container_v1.types.cluster_service import Jwk -from google.cloud.container_v1.types.cluster_service import KubernetesDashboard -from google.cloud.container_v1.types.cluster_service import LegacyAbac -from google.cloud.container_v1.types.cluster_service import LinuxNodeConfig -from google.cloud.container_v1.types.cluster_service import ListClustersRequest -from google.cloud.container_v1.types.cluster_service import ListClustersResponse -from google.cloud.container_v1.types.cluster_service import ListNodePoolsRequest -from google.cloud.container_v1.types.cluster_service import ListNodePoolsResponse -from google.cloud.container_v1.types.cluster_service import ListOperationsRequest -from google.cloud.container_v1.types.cluster_service import ListOperationsResponse -from google.cloud.container_v1.types.cluster_service import ListUsableSubnetworksRequest -from google.cloud.container_v1.types.cluster_service import ( + GcfsConfig, + GcpFilestoreCsiDriverConfig, + GetClusterRequest, + GetJSONWebKeysRequest, + GetJSONWebKeysResponse, + GetNodePoolRequest, + GetOpenIDConfigRequest, + GetOpenIDConfigResponse, + GetOperationRequest, + GetServerConfigRequest, + HorizontalPodAutoscaling, + HttpLoadBalancing, + ILBSubsettingConfig, + IntraNodeVisibilityConfig, + IPAllocationPolicy, + Jwk, + KubernetesDashboard, + LegacyAbac, + LinuxNodeConfig, + ListClustersRequest, + ListClustersResponse, + ListNodePoolsRequest, + ListNodePoolsResponse, + ListOperationsRequest, + ListOperationsResponse, + ListUsableSubnetworksRequest, ListUsableSubnetworksResponse, -) -from google.cloud.container_v1.types.cluster_service import LoggingComponentConfig -from google.cloud.container_v1.types.cluster_service import LoggingConfig -from google.cloud.container_v1.types.cluster_service import MaintenanceExclusionOptions -from google.cloud.container_v1.types.cluster_service import MaintenancePolicy -from google.cloud.container_v1.types.cluster_service import MaintenanceWindow -from google.cloud.container_v1.types.cluster_service import MasterAuth -from google.cloud.container_v1.types.cluster_service import ( + LoggingComponentConfig, + LoggingConfig, + MaintenanceExclusionOptions, + MaintenancePolicy, + MaintenanceWindow, + MasterAuth, MasterAuthorizedNetworksConfig, -) -from google.cloud.container_v1.types.cluster_service import MaxPodsConstraint -from google.cloud.container_v1.types.cluster_service import MeshCertificates -from google.cloud.container_v1.types.cluster_service import MonitoringComponentConfig -from google.cloud.container_v1.types.cluster_service import MonitoringConfig -from google.cloud.container_v1.types.cluster_service import NetworkConfig -from google.cloud.container_v1.types.cluster_service import NetworkPolicy -from google.cloud.container_v1.types.cluster_service import NetworkPolicyConfig -from google.cloud.container_v1.types.cluster_service import NodeConfig -from google.cloud.container_v1.types.cluster_service import NodeConfigDefaults -from google.cloud.container_v1.types.cluster_service import NodeKubeletConfig -from google.cloud.container_v1.types.cluster_service import NodeManagement -from google.cloud.container_v1.types.cluster_service import NodeNetworkConfig -from google.cloud.container_v1.types.cluster_service import NodePool -from google.cloud.container_v1.types.cluster_service import NodePoolAutoscaling -from google.cloud.container_v1.types.cluster_service import NodePoolDefaults -from google.cloud.container_v1.types.cluster_service import NodeTaint -from google.cloud.container_v1.types.cluster_service import NotificationConfig -from google.cloud.container_v1.types.cluster_service import Operation -from google.cloud.container_v1.types.cluster_service import OperationProgress -from google.cloud.container_v1.types.cluster_service import PrivateClusterConfig -from google.cloud.container_v1.types.cluster_service import ( + MaxPodsConstraint, + MeshCertificates, + MonitoringComponentConfig, + MonitoringConfig, + NetworkConfig, + NetworkPolicy, + NetworkPolicyConfig, + NodeConfig, + NodeConfigDefaults, + NodeKubeletConfig, + NodeManagement, + NodeNetworkConfig, + NodePool, + NodePoolAutoscaling, + NodePoolDefaults, + NodeTaint, + NotificationConfig, + Operation, + OperationProgress, + PrivateClusterConfig, PrivateClusterMasterGlobalAccessConfig, -) -from google.cloud.container_v1.types.cluster_service import RecurringTimeWindow -from google.cloud.container_v1.types.cluster_service import ReleaseChannel -from google.cloud.container_v1.types.cluster_service import ReservationAffinity -from google.cloud.container_v1.types.cluster_service import ResourceLimit -from google.cloud.container_v1.types.cluster_service import ResourceUsageExportConfig -from google.cloud.container_v1.types.cluster_service import ( + PrivateIPv6GoogleAccess, + RecurringTimeWindow, + ReleaseChannel, + ReservationAffinity, + ResourceLimit, + ResourceUsageExportConfig, RollbackNodePoolUpgradeRequest, -) -from google.cloud.container_v1.types.cluster_service import SandboxConfig -from google.cloud.container_v1.types.cluster_service import SecurityBulletinEvent -from google.cloud.container_v1.types.cluster_service import ServerConfig -from google.cloud.container_v1.types.cluster_service import ServiceExternalIPsConfig -from google.cloud.container_v1.types.cluster_service import SetAddonsConfigRequest -from google.cloud.container_v1.types.cluster_service import SetLabelsRequest -from google.cloud.container_v1.types.cluster_service import SetLegacyAbacRequest -from google.cloud.container_v1.types.cluster_service import SetLocationsRequest -from google.cloud.container_v1.types.cluster_service import SetLoggingServiceRequest -from google.cloud.container_v1.types.cluster_service import SetMaintenancePolicyRequest -from google.cloud.container_v1.types.cluster_service import SetMasterAuthRequest -from google.cloud.container_v1.types.cluster_service import SetMonitoringServiceRequest -from google.cloud.container_v1.types.cluster_service import SetNetworkPolicyRequest -from google.cloud.container_v1.types.cluster_service import ( + SandboxConfig, + SecurityBulletinEvent, + ServerConfig, + ServiceExternalIPsConfig, + SetAddonsConfigRequest, + SetLabelsRequest, + SetLegacyAbacRequest, + SetLocationsRequest, + SetLoggingServiceRequest, + SetMaintenancePolicyRequest, + SetMasterAuthRequest, + SetMonitoringServiceRequest, + SetNetworkPolicyRequest, SetNodePoolAutoscalingRequest, -) -from google.cloud.container_v1.types.cluster_service import SetNodePoolManagementRequest -from google.cloud.container_v1.types.cluster_service import SetNodePoolSizeRequest -from google.cloud.container_v1.types.cluster_service import ShieldedInstanceConfig -from google.cloud.container_v1.types.cluster_service import ShieldedNodes -from google.cloud.container_v1.types.cluster_service import StartIPRotationRequest -from google.cloud.container_v1.types.cluster_service import StatusCondition -from google.cloud.container_v1.types.cluster_service import TimeWindow -from google.cloud.container_v1.types.cluster_service import UpdateClusterRequest -from google.cloud.container_v1.types.cluster_service import UpdateMasterRequest -from google.cloud.container_v1.types.cluster_service import UpdateNodePoolRequest -from google.cloud.container_v1.types.cluster_service import UpgradeAvailableEvent -from google.cloud.container_v1.types.cluster_service import UpgradeEvent -from google.cloud.container_v1.types.cluster_service import UsableSubnetwork -from google.cloud.container_v1.types.cluster_service import ( + SetNodePoolManagementRequest, + SetNodePoolSizeRequest, + ShieldedInstanceConfig, + ShieldedNodes, + StartIPRotationRequest, + StatusCondition, + TimeWindow, + UpdateClusterRequest, + UpdateMasterRequest, + UpdateNodePoolRequest, + UpgradeAvailableEvent, + UpgradeEvent, + UpgradeResourceType, + UsableSubnetwork, UsableSubnetworkSecondaryRange, + VerticalPodAutoscaling, + VirtualNIC, + WorkloadIdentityConfig, + WorkloadMetadataConfig, ) -from google.cloud.container_v1.types.cluster_service import VerticalPodAutoscaling -from google.cloud.container_v1.types.cluster_service import VirtualNIC -from google.cloud.container_v1.types.cluster_service import WorkloadIdentityConfig -from google.cloud.container_v1.types.cluster_service import WorkloadMetadataConfig -from google.cloud.container_v1.types.cluster_service import DatapathProvider -from google.cloud.container_v1.types.cluster_service import PrivateIPv6GoogleAccess -from google.cloud.container_v1.types.cluster_service import UpgradeResourceType __all__ = ( "ClusterManagerClient", diff --git a/google/cloud/container_v1/__init__.py b/google/cloud/container_v1/__init__.py index 69276ad0..3d33a5e4 100644 --- a/google/cloud/container_v1/__init__.py +++ b/google/cloud/container_v1/__init__.py @@ -14,132 +14,132 @@ # limitations under the License. # -from .services.cluster_manager import ClusterManagerClient -from .services.cluster_manager import ClusterManagerAsyncClient - -from .types.cluster_service import AcceleratorConfig -from .types.cluster_service import AddonsConfig -from .types.cluster_service import AdvancedMachineFeatures -from .types.cluster_service import AuthenticatorGroupsConfig -from .types.cluster_service import Autopilot -from .types.cluster_service import AutoprovisioningNodePoolDefaults -from .types.cluster_service import AutoUpgradeOptions -from .types.cluster_service import BinaryAuthorization -from .types.cluster_service import CancelOperationRequest -from .types.cluster_service import ClientCertificateConfig -from .types.cluster_service import CloudRunConfig -from .types.cluster_service import Cluster -from .types.cluster_service import ClusterAutoscaling -from .types.cluster_service import ClusterUpdate -from .types.cluster_service import CompleteIPRotationRequest -from .types.cluster_service import ConfidentialNodes -from .types.cluster_service import ConfigConnectorConfig -from .types.cluster_service import CreateClusterRequest -from .types.cluster_service import CreateNodePoolRequest -from .types.cluster_service import DailyMaintenanceWindow -from .types.cluster_service import DatabaseEncryption -from .types.cluster_service import DefaultSnatStatus -from .types.cluster_service import DeleteClusterRequest -from .types.cluster_service import DeleteNodePoolRequest -from .types.cluster_service import DnsCacheConfig -from .types.cluster_service import DNSConfig -from .types.cluster_service import GcePersistentDiskCsiDriverConfig -from .types.cluster_service import GcfsConfig -from .types.cluster_service import GcpFilestoreCsiDriverConfig -from .types.cluster_service import GetClusterRequest -from .types.cluster_service import GetJSONWebKeysRequest -from .types.cluster_service import GetJSONWebKeysResponse -from .types.cluster_service import GetNodePoolRequest -from .types.cluster_service import GetOpenIDConfigRequest -from .types.cluster_service import GetOpenIDConfigResponse -from .types.cluster_service import GetOperationRequest -from .types.cluster_service import GetServerConfigRequest -from .types.cluster_service import HorizontalPodAutoscaling -from .types.cluster_service import HttpLoadBalancing -from .types.cluster_service import ILBSubsettingConfig -from .types.cluster_service import IntraNodeVisibilityConfig -from .types.cluster_service import IPAllocationPolicy -from .types.cluster_service import Jwk -from .types.cluster_service import KubernetesDashboard -from .types.cluster_service import LegacyAbac -from .types.cluster_service import LinuxNodeConfig -from .types.cluster_service import ListClustersRequest -from .types.cluster_service import ListClustersResponse -from .types.cluster_service import ListNodePoolsRequest -from .types.cluster_service import ListNodePoolsResponse -from .types.cluster_service import ListOperationsRequest -from .types.cluster_service import ListOperationsResponse -from .types.cluster_service import ListUsableSubnetworksRequest -from .types.cluster_service import ListUsableSubnetworksResponse -from .types.cluster_service import LoggingComponentConfig -from .types.cluster_service import LoggingConfig -from .types.cluster_service import MaintenanceExclusionOptions -from .types.cluster_service import MaintenancePolicy -from .types.cluster_service import MaintenanceWindow -from .types.cluster_service import MasterAuth -from .types.cluster_service import MasterAuthorizedNetworksConfig -from .types.cluster_service import MaxPodsConstraint -from .types.cluster_service import MeshCertificates -from .types.cluster_service import MonitoringComponentConfig -from .types.cluster_service import MonitoringConfig -from .types.cluster_service import NetworkConfig -from .types.cluster_service import NetworkPolicy -from .types.cluster_service import NetworkPolicyConfig -from .types.cluster_service import NodeConfig -from .types.cluster_service import NodeConfigDefaults -from .types.cluster_service import NodeKubeletConfig -from .types.cluster_service import NodeManagement -from .types.cluster_service import NodeNetworkConfig -from .types.cluster_service import NodePool -from .types.cluster_service import NodePoolAutoscaling -from .types.cluster_service import NodePoolDefaults -from .types.cluster_service import NodeTaint -from .types.cluster_service import NotificationConfig -from .types.cluster_service import Operation -from .types.cluster_service import OperationProgress -from .types.cluster_service import PrivateClusterConfig -from .types.cluster_service import PrivateClusterMasterGlobalAccessConfig -from .types.cluster_service import RecurringTimeWindow -from .types.cluster_service import ReleaseChannel -from .types.cluster_service import ReservationAffinity -from .types.cluster_service import ResourceLimit -from .types.cluster_service import ResourceUsageExportConfig -from .types.cluster_service import RollbackNodePoolUpgradeRequest -from .types.cluster_service import SandboxConfig -from .types.cluster_service import SecurityBulletinEvent -from .types.cluster_service import ServerConfig -from .types.cluster_service import ServiceExternalIPsConfig -from .types.cluster_service import SetAddonsConfigRequest -from .types.cluster_service import SetLabelsRequest -from .types.cluster_service import SetLegacyAbacRequest -from .types.cluster_service import SetLocationsRequest -from .types.cluster_service import SetLoggingServiceRequest -from .types.cluster_service import SetMaintenancePolicyRequest -from .types.cluster_service import SetMasterAuthRequest -from .types.cluster_service import SetMonitoringServiceRequest -from .types.cluster_service import SetNetworkPolicyRequest -from .types.cluster_service import SetNodePoolAutoscalingRequest -from .types.cluster_service import SetNodePoolManagementRequest -from .types.cluster_service import SetNodePoolSizeRequest -from .types.cluster_service import ShieldedInstanceConfig -from .types.cluster_service import ShieldedNodes -from .types.cluster_service import StartIPRotationRequest -from .types.cluster_service import StatusCondition -from .types.cluster_service import TimeWindow -from .types.cluster_service import UpdateClusterRequest -from .types.cluster_service import UpdateMasterRequest -from .types.cluster_service import UpdateNodePoolRequest -from .types.cluster_service import UpgradeAvailableEvent -from .types.cluster_service import UpgradeEvent -from .types.cluster_service import UsableSubnetwork -from .types.cluster_service import UsableSubnetworkSecondaryRange -from .types.cluster_service import VerticalPodAutoscaling -from .types.cluster_service import VirtualNIC -from .types.cluster_service import WorkloadIdentityConfig -from .types.cluster_service import WorkloadMetadataConfig -from .types.cluster_service import DatapathProvider -from .types.cluster_service import PrivateIPv6GoogleAccess -from .types.cluster_service import UpgradeResourceType +from .services.cluster_manager import ClusterManagerAsyncClient, ClusterManagerClient +from .types.cluster_service import ( + AcceleratorConfig, + AddonsConfig, + AdvancedMachineFeatures, + AuthenticatorGroupsConfig, + Autopilot, + AutoprovisioningNodePoolDefaults, + AutoUpgradeOptions, + BinaryAuthorization, + CancelOperationRequest, + ClientCertificateConfig, + CloudRunConfig, + Cluster, + ClusterAutoscaling, + ClusterUpdate, + CompleteIPRotationRequest, + ConfidentialNodes, + ConfigConnectorConfig, + CreateClusterRequest, + CreateNodePoolRequest, + DailyMaintenanceWindow, + DatabaseEncryption, + DatapathProvider, + DefaultSnatStatus, + DeleteClusterRequest, + DeleteNodePoolRequest, + DnsCacheConfig, + DNSConfig, + GcePersistentDiskCsiDriverConfig, + GcfsConfig, + GcpFilestoreCsiDriverConfig, + GetClusterRequest, + GetJSONWebKeysRequest, + GetJSONWebKeysResponse, + GetNodePoolRequest, + GetOpenIDConfigRequest, + GetOpenIDConfigResponse, + GetOperationRequest, + GetServerConfigRequest, + HorizontalPodAutoscaling, + HttpLoadBalancing, + ILBSubsettingConfig, + IntraNodeVisibilityConfig, + IPAllocationPolicy, + Jwk, + KubernetesDashboard, + LegacyAbac, + LinuxNodeConfig, + ListClustersRequest, + ListClustersResponse, + ListNodePoolsRequest, + ListNodePoolsResponse, + ListOperationsRequest, + ListOperationsResponse, + ListUsableSubnetworksRequest, + ListUsableSubnetworksResponse, + LoggingComponentConfig, + LoggingConfig, + MaintenanceExclusionOptions, + MaintenancePolicy, + MaintenanceWindow, + MasterAuth, + MasterAuthorizedNetworksConfig, + MaxPodsConstraint, + MeshCertificates, + MonitoringComponentConfig, + MonitoringConfig, + NetworkConfig, + NetworkPolicy, + NetworkPolicyConfig, + NodeConfig, + NodeConfigDefaults, + NodeKubeletConfig, + NodeManagement, + NodeNetworkConfig, + NodePool, + NodePoolAutoscaling, + NodePoolDefaults, + NodeTaint, + NotificationConfig, + Operation, + OperationProgress, + PrivateClusterConfig, + PrivateClusterMasterGlobalAccessConfig, + PrivateIPv6GoogleAccess, + RecurringTimeWindow, + ReleaseChannel, + ReservationAffinity, + ResourceLimit, + ResourceUsageExportConfig, + RollbackNodePoolUpgradeRequest, + SandboxConfig, + SecurityBulletinEvent, + ServerConfig, + ServiceExternalIPsConfig, + SetAddonsConfigRequest, + SetLabelsRequest, + SetLegacyAbacRequest, + SetLocationsRequest, + SetLoggingServiceRequest, + SetMaintenancePolicyRequest, + SetMasterAuthRequest, + SetMonitoringServiceRequest, + SetNetworkPolicyRequest, + SetNodePoolAutoscalingRequest, + SetNodePoolManagementRequest, + SetNodePoolSizeRequest, + ShieldedInstanceConfig, + ShieldedNodes, + StartIPRotationRequest, + StatusCondition, + TimeWindow, + UpdateClusterRequest, + UpdateMasterRequest, + UpdateNodePoolRequest, + UpgradeAvailableEvent, + UpgradeEvent, + UpgradeResourceType, + UsableSubnetwork, + UsableSubnetworkSecondaryRange, + VerticalPodAutoscaling, + VirtualNIC, + WorkloadIdentityConfig, + WorkloadMetadataConfig, +) __all__ = ( "ClusterManagerAsyncClient", diff --git a/google/cloud/container_v1/services/cluster_manager/__init__.py b/google/cloud/container_v1/services/cluster_manager/__init__.py index da06b070..8445ce15 100644 --- a/google/cloud/container_v1/services/cluster_manager/__init__.py +++ b/google/cloud/container_v1/services/cluster_manager/__init__.py @@ -13,8 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # -from .client import ClusterManagerClient from .async_client import ClusterManagerAsyncClient +from .client import ClusterManagerClient __all__ = ( "ClusterManagerClient", diff --git a/google/cloud/container_v1/services/cluster_manager/async_client.py b/google/cloud/container_v1/services/cluster_manager/async_client.py index df13bdbc..ac397bcd 100644 --- a/google/cloud/container_v1/services/cluster_manager/async_client.py +++ b/google/cloud/container_v1/services/cluster_manager/async_client.py @@ -16,28 +16,30 @@ from collections import OrderedDict import functools import re -from typing import Dict, Optional, Sequence, Tuple, Type, Union -import pkg_resources +from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union import warnings -from google.api_core.client_options import ClientOptions from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 from google.api_core import retry as retries +from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore +import pkg_resources try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] except AttributeError: # pragma: NO COVER OptionalRetry = Union[retries.Retry, object] # type: ignore +from google.rpc import status_pb2 # type: ignore + from google.cloud.container_v1.services.cluster_manager import pagers from google.cloud.container_v1.types import cluster_service -from google.rpc import status_pb2 # type: ignore -from .transports.base import ClusterManagerTransport, DEFAULT_CLIENT_INFO -from .transports.grpc_asyncio import ClusterManagerGrpcAsyncIOTransport + from .client import ClusterManagerClient +from .transports.base import DEFAULT_CLIENT_INFO, ClusterManagerTransport +from .transports.grpc_asyncio import ClusterManagerGrpcAsyncIOTransport class ClusterManagerAsyncClient: @@ -218,21 +220,20 @@ async def list_clusters( r"""Lists all clusters owned by a project in either the specified zone or all zones. - .. code-block:: python from google.cloud import container_v1 - def sample_list_clusters(): + async def sample_list_clusters(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.ListClustersRequest( ) # Make the request - response = client.list_clusters(request=request) + response = await client.list_clusters(request=request) # Handle the response print(response) @@ -327,7 +328,12 @@ def sample_list_clusters(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -350,16 +356,16 @@ async def get_cluster( from google.cloud import container_v1 - def sample_get_cluster(): + async def sample_get_cluster(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.GetClusterRequest( ) # Make the request - response = client.get_cluster(request=request) + response = await client.get_cluster(request=request) # Handle the response print(response) @@ -462,7 +468,12 @@ def sample_get_cluster(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -493,21 +504,20 @@ async def create_cluster( Finally, an entry is added to the project's global metadata indicating which CIDR range the cluster is using. - .. code-block:: python from google.cloud import container_v1 - def sample_create_cluster(): + async def sample_create_cluster(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.CreateClusterRequest( ) # Make the request - response = client.create_cluster(request=request) + response = await client.create_cluster(request=request) # Handle the response print(response) @@ -602,7 +612,12 @@ def sample_create_cluster(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -626,16 +641,16 @@ async def update_cluster( from google.cloud import container_v1 - def sample_update_cluster(): + async def sample_update_cluster(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.UpdateClusterRequest( ) # Make the request - response = client.update_cluster(request=request) + response = await client.update_cluster(request=request) # Handle the response print(response) @@ -741,7 +756,12 @@ def sample_update_cluster(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -757,14 +777,13 @@ async def update_node_pool( r"""Updates the version and/or image type for the specified node pool. - .. code-block:: python from google.cloud import container_v1 - def sample_update_node_pool(): + async def sample_update_node_pool(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.UpdateNodePoolRequest( @@ -773,7 +792,7 @@ def sample_update_node_pool(): ) # Make the request - response = client.update_node_pool(request=request) + response = await client.update_node_pool(request=request) # Handle the response print(response) @@ -814,7 +833,12 @@ def sample_update_node_pool(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -830,21 +854,20 @@ async def set_node_pool_autoscaling( r"""Sets the autoscaling settings for the specified node pool. - .. code-block:: python from google.cloud import container_v1 - def sample_set_node_pool_autoscaling(): + async def sample_set_node_pool_autoscaling(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetNodePoolAutoscalingRequest( ) # Make the request - response = client.set_node_pool_autoscaling(request=request) + response = await client.set_node_pool_autoscaling(request=request) # Handle the response print(response) @@ -885,7 +908,12 @@ def sample_set_node_pool_autoscaling(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -909,9 +937,9 @@ async def set_logging_service( from google.cloud import container_v1 - def sample_set_logging_service(): + async def sample_set_logging_service(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetLoggingServiceRequest( @@ -919,7 +947,7 @@ def sample_set_logging_service(): ) # Make the request - response = client.set_logging_service(request=request) + response = await client.set_logging_service(request=request) # Handle the response print(response) @@ -1039,7 +1067,12 @@ def sample_set_logging_service(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1063,9 +1096,9 @@ async def set_monitoring_service( from google.cloud import container_v1 - def sample_set_monitoring_service(): + async def sample_set_monitoring_service(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetMonitoringServiceRequest( @@ -1073,7 +1106,7 @@ def sample_set_monitoring_service(): ) # Make the request - response = client.set_monitoring_service(request=request) + response = await client.set_monitoring_service(request=request) # Handle the response print(response) @@ -1195,7 +1228,12 @@ def sample_set_monitoring_service(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1219,16 +1257,16 @@ async def set_addons_config( from google.cloud import container_v1 - def sample_set_addons_config(): + async def sample_set_addons_config(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetAddonsConfigRequest( ) # Make the request - response = client.set_addons_config(request=request) + response = await client.set_addons_config(request=request) # Handle the response print(response) @@ -1335,7 +1373,12 @@ def sample_set_addons_config(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1357,14 +1400,13 @@ async def set_locations( `projects.locations.clusters.update `__ instead. - .. code-block:: python from google.cloud import container_v1 - def sample_set_locations(): + async def sample_set_locations(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetLocationsRequest( @@ -1372,7 +1414,7 @@ def sample_set_locations(): ) # Make the request - response = client.set_locations(request=request) + response = await client.set_locations(request=request) # Handle the response print(response) @@ -1489,7 +1531,12 @@ def sample_set_locations(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1513,9 +1560,9 @@ async def update_master( from google.cloud import container_v1 - def sample_update_master(): + async def sample_update_master(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.UpdateMasterRequest( @@ -1523,7 +1570,7 @@ def sample_update_master(): ) # Make the request - response = client.update_master(request=request) + response = await client.update_master(request=request) # Handle the response print(response) @@ -1641,7 +1688,12 @@ def sample_update_master(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1659,14 +1711,13 @@ async def set_master_auth( either via password generation or explicitly setting the password. - .. code-block:: python from google.cloud import container_v1 - def sample_set_master_auth(): + async def sample_set_master_auth(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetMasterAuthRequest( @@ -1674,7 +1725,7 @@ def sample_set_master_auth(): ) # Make the request - response = client.set_master_auth(request=request) + response = await client.set_master_auth(request=request) # Handle the response print(response) @@ -1715,7 +1766,12 @@ def sample_set_master_auth(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1743,21 +1799,20 @@ async def delete_cluster( not deleted if they weren't present when the cluster was initially created. - .. code-block:: python from google.cloud import container_v1 - def sample_delete_cluster(): + async def sample_delete_cluster(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.DeleteClusterRequest( ) # Make the request - response = client.delete_cluster(request=request) + response = await client.delete_cluster(request=request) # Handle the response print(response) @@ -1864,7 +1919,12 @@ def sample_delete_cluster(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1882,21 +1942,20 @@ async def list_operations( r"""Lists all operations in a project in a specific zone or all zones. - .. code-block:: python from google.cloud import container_v1 - def sample_list_operations(): + async def sample_list_operations(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.ListOperationsRequest( ) # Make the request - response = client.list_operations(request=request) + response = await client.list_operations(request=request) # Handle the response print(response) @@ -1981,7 +2040,12 @@ def sample_list_operations(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2004,16 +2068,16 @@ async def get_operation( from google.cloud import container_v1 - def sample_get_operation(): + async def sample_get_operation(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.GetOperationRequest( ) # Make the request - response = client.get_operation(request=request) + response = await client.get_operation(request=request) # Handle the response print(response) @@ -2119,7 +2183,12 @@ def sample_get_operation(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2142,16 +2211,16 @@ async def cancel_operation( from google.cloud import container_v1 - def sample_cancel_operation(): + async def sample_cancel_operation(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.CancelOperationRequest( ) # Make the request - client.cancel_operation(request=request) + await client.cancel_operation(request=request) Args: request (Union[google.cloud.container_v1.types.CancelOperationRequest, dict]): @@ -2237,7 +2306,10 @@ def sample_cancel_operation(): # Send the request. await rpc( - request, retry=retry, timeout=timeout, metadata=metadata, + request, + retry=retry, + timeout=timeout, + metadata=metadata, ) async def get_server_config( @@ -2254,21 +2326,20 @@ async def get_server_config( r"""Returns configuration info about the Google Kubernetes Engine service. - .. code-block:: python from google.cloud import container_v1 - def sample_get_server_config(): + async def sample_get_server_config(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.GetServerConfigRequest( ) # Make the request - response = client.get_server_config(request=request) + response = await client.get_server_config(request=request) # Handle the response print(response) @@ -2361,7 +2432,12 @@ def sample_get_server_config(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2379,21 +2455,20 @@ async def get_json_web_keys( This API is not yet intended for general use, and is not available for all clusters. - .. code-block:: python from google.cloud import container_v1 - def sample_get_json_web_keys(): + async def sample_get_json_web_keys(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.GetJSONWebKeysRequest( ) # Make the request - response = client.get_json_web_keys(request=request) + response = await client.get_json_web_keys(request=request) # Handle the response print(response) @@ -2436,7 +2511,12 @@ def sample_get_json_web_keys(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2459,16 +2539,16 @@ async def list_node_pools( from google.cloud import container_v1 - def sample_list_node_pools(): + async def sample_list_node_pools(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.ListNodePoolsRequest( ) # Make the request - response = client.list_node_pools(request=request) + response = await client.list_node_pools(request=request) # Handle the response print(response) @@ -2572,7 +2652,12 @@ def sample_list_node_pools(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2596,16 +2681,16 @@ async def get_node_pool( from google.cloud import container_v1 - def sample_get_node_pool(): + async def sample_get_node_pool(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.GetNodePoolRequest( ) # Make the request - response = client.get_node_pool(request=request) + response = await client.get_node_pool(request=request) # Handle the response print(response) @@ -2727,7 +2812,12 @@ def sample_get_node_pool(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2751,16 +2841,16 @@ async def create_node_pool( from google.cloud import container_v1 - def sample_create_node_pool(): + async def sample_create_node_pool(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.CreateNodePoolRequest( ) # Make the request - response = client.create_node_pool(request=request) + response = await client.create_node_pool(request=request) # Handle the response print(response) @@ -2863,7 +2953,12 @@ def sample_create_node_pool(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2887,16 +2982,16 @@ async def delete_node_pool( from google.cloud import container_v1 - def sample_delete_node_pool(): + async def sample_delete_node_pool(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.DeleteNodePoolRequest( ) # Make the request - response = client.delete_node_pool(request=request) + response = await client.delete_node_pool(request=request) # Handle the response print(response) @@ -3013,7 +3108,12 @@ def sample_delete_node_pool(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3035,21 +3135,20 @@ async def rollback_node_pool_upgrade( upgrade. This makes no changes if the last upgrade successfully completed. - .. code-block:: python from google.cloud import container_v1 - def sample_rollback_node_pool_upgrade(): + async def sample_rollback_node_pool_upgrade(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.RollbackNodePoolUpgradeRequest( ) # Make the request - response = client.rollback_node_pool_upgrade(request=request) + response = await client.rollback_node_pool_upgrade(request=request) # Handle the response print(response) @@ -3160,7 +3259,12 @@ def sample_rollback_node_pool_upgrade(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3179,16 +3283,16 @@ async def set_node_pool_management( from google.cloud import container_v1 - def sample_set_node_pool_management(): + async def sample_set_node_pool_management(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetNodePoolManagementRequest( ) # Make the request - response = client.set_node_pool_management(request=request) + response = await client.set_node_pool_management(request=request) # Handle the response print(response) @@ -3229,7 +3333,12 @@ def sample_set_node_pool_management(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3248,9 +3357,9 @@ async def set_labels( from google.cloud import container_v1 - def sample_set_labels(): + async def sample_set_labels(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetLabelsRequest( @@ -3258,7 +3367,7 @@ def sample_set_labels(): ) # Make the request - response = client.set_labels(request=request) + response = await client.set_labels(request=request) # Handle the response print(response) @@ -3301,7 +3410,12 @@ def sample_set_labels(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3322,14 +3436,13 @@ async def set_legacy_abac( r"""Enables or disables the ABAC authorization mechanism on a cluster. - .. code-block:: python from google.cloud import container_v1 - def sample_set_legacy_abac(): + async def sample_set_legacy_abac(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetLegacyAbacRequest( @@ -3337,7 +3450,7 @@ def sample_set_legacy_abac(): ) # Make the request - response = client.set_legacy_abac(request=request) + response = await client.set_legacy_abac(request=request) # Handle the response print(response) @@ -3443,7 +3556,12 @@ def sample_set_legacy_abac(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3466,16 +3584,16 @@ async def start_ip_rotation( from google.cloud import container_v1 - def sample_start_ip_rotation(): + async def sample_start_ip_rotation(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.StartIPRotationRequest( ) # Make the request - response = client.start_ip_rotation(request=request) + response = await client.start_ip_rotation(request=request) # Handle the response print(response) @@ -3572,7 +3690,12 @@ def sample_start_ip_rotation(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3595,16 +3718,16 @@ async def complete_ip_rotation( from google.cloud import container_v1 - def sample_complete_ip_rotation(): + async def sample_complete_ip_rotation(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.CompleteIPRotationRequest( ) # Make the request - response = client.complete_ip_rotation(request=request) + response = await client.complete_ip_rotation(request=request) # Handle the response print(response) @@ -3700,7 +3823,12 @@ def sample_complete_ip_rotation(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3718,14 +3846,13 @@ async def set_node_pool_size( modifying [NodePool.locations][google.container.v1.NodePool.locations]. - .. code-block:: python from google.cloud import container_v1 - def sample_set_node_pool_size(): + async def sample_set_node_pool_size(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetNodePoolSizeRequest( @@ -3733,7 +3860,7 @@ def sample_set_node_pool_size(): ) # Make the request - response = client.set_node_pool_size(request=request) + response = await client.set_node_pool_size(request=request) # Handle the response print(response) @@ -3774,7 +3901,12 @@ def sample_set_node_pool_size(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3798,16 +3930,16 @@ async def set_network_policy( from google.cloud import container_v1 - def sample_set_network_policy(): + async def sample_set_network_policy(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetNetworkPolicyRequest( ) # Make the request - response = client.set_network_policy(request=request) + response = await client.set_network_policy(request=request) # Handle the response print(response) @@ -3912,7 +4044,12 @@ def sample_set_network_policy(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3936,9 +4073,9 @@ async def set_maintenance_policy( from google.cloud import container_v1 - def sample_set_maintenance_policy(): + async def sample_set_maintenance_policy(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.SetMaintenancePolicyRequest( @@ -3948,7 +4085,7 @@ def sample_set_maintenance_policy(): ) # Make the request - response = client.set_maintenance_policy(request=request) + response = await client.set_maintenance_policy(request=request) # Handle the response print(response) @@ -4052,7 +4189,12 @@ def sample_set_maintenance_policy(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -4068,14 +4210,13 @@ async def list_usable_subnetworks( r"""Lists subnetworks that are usable for creating clusters in a project. - .. code-block:: python from google.cloud import container_v1 - def sample_list_usable_subnetworks(): + async def sample_list_usable_subnetworks(): # Create a client - client = container_v1.ClusterManagerClient() + client = container_v1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1.ListUsableSubnetworksRequest( @@ -4085,7 +4226,7 @@ def sample_list_usable_subnetworks(): page_result = client.list_usable_subnetworks(request=request) # Handle the response - for response in page_result: + async for response in page_result: print(response) Args: @@ -4128,12 +4269,20 @@ def sample_list_usable_subnetworks(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__aiter__` convenience method. response = pagers.ListUsableSubnetworksAsyncPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -4148,7 +4297,9 @@ async def __aexit__(self, exc_type, exc, tb): try: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution("google-cloud-container",).version, + gapic_version=pkg_resources.get_distribution( + "google-cloud-container", + ).version, ) except pkg_resources.DistributionNotFound: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() diff --git a/google/cloud/container_v1/services/cluster_manager/client.py b/google/cloud/container_v1/services/cluster_manager/client.py index 1faf11c8..4f7bf894 100644 --- a/google/cloud/container_v1/services/cluster_manager/client.py +++ b/google/cloud/container_v1/services/cluster_manager/client.py @@ -16,8 +16,7 @@ from collections import OrderedDict import os import re -from typing import Dict, Optional, Sequence, Tuple, Type, Union -import pkg_resources +from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union import warnings from google.api_core import client_options as client_options_lib @@ -25,20 +24,23 @@ from google.api_core import gapic_v1 from google.api_core import retry as retries from google.auth import credentials as ga_credentials # type: ignore +from google.auth.exceptions import MutualTLSChannelError # type: ignore from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore -from google.auth.exceptions import MutualTLSChannelError # type: ignore from google.oauth2 import service_account # type: ignore +import pkg_resources try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] except AttributeError: # pragma: NO COVER OptionalRetry = Union[retries.Retry, object] # type: ignore +from google.rpc import status_pb2 # type: ignore + from google.cloud.container_v1.services.cluster_manager import pagers from google.cloud.container_v1.types import cluster_service -from google.rpc import status_pb2 # type: ignore -from .transports.base import ClusterManagerTransport, DEFAULT_CLIENT_INFO + +from .transports.base import DEFAULT_CLIENT_INFO, ClusterManagerTransport from .transports.grpc import ClusterManagerGrpcTransport from .transports.grpc_asyncio import ClusterManagerGrpcAsyncIOTransport @@ -57,7 +59,10 @@ class ClusterManagerClientMeta(type): _transport_registry["grpc"] = ClusterManagerGrpcTransport _transport_registry["grpc_asyncio"] = ClusterManagerGrpcAsyncIOTransport - def get_transport_class(cls, label: str = None,) -> Type[ClusterManagerTransport]: + def get_transport_class( + cls, + label: str = None, + ) -> Type[ClusterManagerTransport]: """Returns an appropriate transport class. Args: @@ -162,9 +167,15 @@ def transport(self) -> ClusterManagerTransport: return self._transport @staticmethod - def topic_path(project: str, topic: str,) -> str: + def topic_path( + project: str, + topic: str, + ) -> str: """Returns a fully-qualified topic string.""" - return "projects/{project}/topics/{topic}".format(project=project, topic=topic,) + return "projects/{project}/topics/{topic}".format( + project=project, + topic=topic, + ) @staticmethod def parse_topic_path(path: str) -> Dict[str, str]: @@ -173,7 +184,9 @@ def parse_topic_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_billing_account_path(billing_account: str,) -> str: + def common_billing_account_path( + billing_account: str, + ) -> str: """Returns a fully-qualified billing_account string.""" return "billingAccounts/{billing_account}".format( billing_account=billing_account, @@ -186,9 +199,13 @@ def parse_common_billing_account_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_folder_path(folder: str,) -> str: + def common_folder_path( + folder: str, + ) -> str: """Returns a fully-qualified folder string.""" - return "folders/{folder}".format(folder=folder,) + return "folders/{folder}".format( + folder=folder, + ) @staticmethod def parse_common_folder_path(path: str) -> Dict[str, str]: @@ -197,9 +214,13 @@ def parse_common_folder_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_organization_path(organization: str,) -> str: + def common_organization_path( + organization: str, + ) -> str: """Returns a fully-qualified organization string.""" - return "organizations/{organization}".format(organization=organization,) + return "organizations/{organization}".format( + organization=organization, + ) @staticmethod def parse_common_organization_path(path: str) -> Dict[str, str]: @@ -208,9 +229,13 @@ def parse_common_organization_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_project_path(project: str,) -> str: + def common_project_path( + project: str, + ) -> str: """Returns a fully-qualified project string.""" - return "projects/{project}".format(project=project,) + return "projects/{project}".format( + project=project, + ) @staticmethod def parse_common_project_path(path: str) -> Dict[str, str]: @@ -219,10 +244,14 @@ def parse_common_project_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_location_path(project: str, location: str,) -> str: + def common_location_path( + project: str, + location: str, + ) -> str: """Returns a fully-qualified location string.""" return "projects/{project}/locations/{location}".format( - project=project, location=location, + project=project, + location=location, ) @staticmethod @@ -410,7 +439,6 @@ def list_clusters( r"""Lists all clusters owned by a project in either the specified zone or all zones. - .. code-block:: python from google.cloud import container_v1 @@ -509,7 +537,12 @@ def sample_list_clusters(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -634,7 +667,12 @@ def sample_get_cluster(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -665,7 +703,6 @@ def create_cluster( Finally, an entry is added to the project's global metadata indicating which CIDR range the cluster is using. - .. code-block:: python from google.cloud import container_v1 @@ -774,7 +811,12 @@ def sample_create_cluster(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -913,7 +955,12 @@ def sample_update_cluster(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -929,7 +976,6 @@ def update_node_pool( r"""Updates the version and/or image type for the specified node pool. - .. code-block:: python from google.cloud import container_v1 @@ -987,7 +1033,12 @@ def sample_update_node_pool(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1003,7 +1054,6 @@ def set_node_pool_autoscaling( r"""Sets the autoscaling settings for the specified node pool. - .. code-block:: python from google.cloud import container_v1 @@ -1061,7 +1111,12 @@ def sample_set_node_pool_autoscaling(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1215,7 +1270,12 @@ def sample_set_logging_service(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1371,7 +1431,12 @@ def sample_set_monitoring_service(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1511,7 +1576,12 @@ def sample_set_addons_config(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1533,7 +1603,6 @@ def set_locations( `projects.locations.clusters.update `__ instead. - .. code-block:: python from google.cloud import container_v1 @@ -1665,7 +1734,12 @@ def sample_set_locations(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1817,7 +1891,12 @@ def sample_update_master(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1835,7 +1914,6 @@ def set_master_auth( either via password generation or explicitly setting the password. - .. code-block:: python from google.cloud import container_v1 @@ -1892,7 +1970,12 @@ def sample_set_master_auth(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1920,7 +2003,6 @@ def delete_cluster( not deleted if they weren't present when the cluster was initially created. - .. code-block:: python from google.cloud import container_v1 @@ -2031,7 +2113,12 @@ def sample_delete_cluster(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2049,7 +2136,6 @@ def list_operations( r"""Lists all operations in a project in a specific zone or all zones. - .. code-block:: python from google.cloud import container_v1 @@ -2138,7 +2224,12 @@ def sample_list_operations(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2266,7 +2357,12 @@ def sample_get_operation(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2384,7 +2480,10 @@ def sample_cancel_operation(): # Send the request. rpc( - request, retry=retry, timeout=timeout, metadata=metadata, + request, + retry=retry, + timeout=timeout, + metadata=metadata, ) def get_server_config( @@ -2401,7 +2500,6 @@ def get_server_config( r"""Returns configuration info about the Google Kubernetes Engine service. - .. code-block:: python from google.cloud import container_v1 @@ -2498,7 +2596,12 @@ def sample_get_server_config(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2516,7 +2619,6 @@ def get_json_web_keys( This API is not yet intended for general use, and is not available for all clusters. - .. code-block:: python from google.cloud import container_v1 @@ -2574,7 +2676,12 @@ def sample_get_json_web_keys(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2700,7 +2807,12 @@ def sample_list_node_pools(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2845,7 +2957,12 @@ def sample_get_node_pool(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2981,7 +3098,12 @@ def sample_create_node_pool(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3121,7 +3243,12 @@ def sample_delete_node_pool(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3143,7 +3270,6 @@ def rollback_node_pool_upgrade( upgrade. This makes no changes if the last upgrade successfully completed. - .. code-block:: python from google.cloud import container_v1 @@ -3270,7 +3396,12 @@ def sample_rollback_node_pool_upgrade(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3340,7 +3471,12 @@ def sample_set_node_pool_management(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3413,7 +3549,12 @@ def sample_set_labels(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3434,7 +3575,6 @@ def set_legacy_abac( r"""Enables or disables the ABAC authorization mechanism on a cluster. - .. code-block:: python from google.cloud import container_v1 @@ -3555,7 +3695,12 @@ def sample_set_legacy_abac(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3684,7 +3829,12 @@ def sample_start_ip_rotation(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3812,7 +3962,12 @@ def sample_complete_ip_rotation(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3830,7 +3985,6 @@ def set_node_pool_size( modifying [NodePool.locations][google.container.v1.NodePool.locations]. - .. code-block:: python from google.cloud import container_v1 @@ -3887,7 +4041,12 @@ def sample_set_node_pool_size(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -4025,7 +4184,12 @@ def sample_set_network_policy(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -4165,7 +4329,12 @@ def sample_set_maintenance_policy(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -4181,7 +4350,6 @@ def list_usable_subnetworks( r"""Lists subnetworks that are usable for creating clusters in a project. - .. code-block:: python from google.cloud import container_v1 @@ -4242,12 +4410,20 @@ def sample_list_usable_subnetworks(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__iter__` convenience method. response = pagers.ListUsableSubnetworksPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -4269,7 +4445,9 @@ def __exit__(self, type, value, traceback): try: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution("google-cloud-container",).version, + gapic_version=pkg_resources.get_distribution( + "google-cloud-container", + ).version, ) except pkg_resources.DistributionNotFound: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() diff --git a/google/cloud/container_v1/services/cluster_manager/pagers.py b/google/cloud/container_v1/services/cluster_manager/pagers.py index dc606db3..39f4b7c8 100644 --- a/google/cloud/container_v1/services/cluster_manager/pagers.py +++ b/google/cloud/container_v1/services/cluster_manager/pagers.py @@ -18,10 +18,10 @@ AsyncIterator, Awaitable, Callable, + Iterator, + Optional, Sequence, Tuple, - Optional, - Iterator, ) from google.cloud.container_v1.types import cluster_service diff --git a/google/cloud/container_v1/services/cluster_manager/transports/__init__.py b/google/cloud/container_v1/services/cluster_manager/transports/__init__.py index b6375e92..b794d72c 100644 --- a/google/cloud/container_v1/services/cluster_manager/transports/__init__.py +++ b/google/cloud/container_v1/services/cluster_manager/transports/__init__.py @@ -20,7 +20,6 @@ from .grpc import ClusterManagerGrpcTransport from .grpc_asyncio import ClusterManagerGrpcAsyncIOTransport - # Compile a registry of transports. _transport_registry = OrderedDict() # type: Dict[str, Type[ClusterManagerTransport]] _transport_registry["grpc"] = ClusterManagerGrpcTransport diff --git a/google/cloud/container_v1/services/cluster_manager/transports/base.py b/google/cloud/container_v1/services/cluster_manager/transports/base.py index f8dc724a..fbcaad68 100644 --- a/google/cloud/container_v1/services/cluster_manager/transports/base.py +++ b/google/cloud/container_v1/services/cluster_manager/transports/base.py @@ -15,22 +15,24 @@ # import abc from typing import Awaitable, Callable, Dict, Optional, Sequence, Union -import pkg_resources -import google.auth # type: ignore import google.api_core from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 from google.api_core import retry as retries +import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore +from google.protobuf import empty_pb2 # type: ignore +import pkg_resources from google.cloud.container_v1.types import cluster_service -from google.protobuf import empty_pb2 # type: ignore try: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution("google-cloud-container",).version, + gapic_version=pkg_resources.get_distribution( + "google-cloud-container", + ).version, ) except pkg_resources.DistributionNotFound: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() @@ -79,6 +81,7 @@ def __init__( always_use_jwt_access (Optional[bool]): Whether self signed JWT should be used for service account credentials. """ + # Save the hostname. Default to port 443 (HTTPS) if none is specified. if ":" not in host: host += ":443" @@ -150,13 +153,19 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.create_cluster: gapic_v1.method.wrap_method( - self.create_cluster, default_timeout=45.0, client_info=client_info, + self.create_cluster, + default_timeout=45.0, + client_info=client_info, ), self.update_cluster: gapic_v1.method.wrap_method( - self.update_cluster, default_timeout=45.0, client_info=client_info, + self.update_cluster, + default_timeout=45.0, + client_info=client_info, ), self.update_node_pool: gapic_v1.method.wrap_method( - self.update_node_pool, default_timeout=45.0, client_info=client_info, + self.update_node_pool, + default_timeout=45.0, + client_info=client_info, ), self.set_node_pool_autoscaling: gapic_v1.method.wrap_method( self.set_node_pool_autoscaling, @@ -164,7 +173,9 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.set_logging_service: gapic_v1.method.wrap_method( - self.set_logging_service, default_timeout=45.0, client_info=client_info, + self.set_logging_service, + default_timeout=45.0, + client_info=client_info, ), self.set_monitoring_service: gapic_v1.method.wrap_method( self.set_monitoring_service, @@ -172,16 +183,24 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.set_addons_config: gapic_v1.method.wrap_method( - self.set_addons_config, default_timeout=45.0, client_info=client_info, + self.set_addons_config, + default_timeout=45.0, + client_info=client_info, ), self.set_locations: gapic_v1.method.wrap_method( - self.set_locations, default_timeout=45.0, client_info=client_info, + self.set_locations, + default_timeout=45.0, + client_info=client_info, ), self.update_master: gapic_v1.method.wrap_method( - self.update_master, default_timeout=45.0, client_info=client_info, + self.update_master, + default_timeout=45.0, + client_info=client_info, ), self.set_master_auth: gapic_v1.method.wrap_method( - self.set_master_auth, default_timeout=45.0, client_info=client_info, + self.set_master_auth, + default_timeout=45.0, + client_info=client_info, ), self.delete_cluster: gapic_v1.method.wrap_method( self.delete_cluster, @@ -229,7 +248,9 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.cancel_operation: gapic_v1.method.wrap_method( - self.cancel_operation, default_timeout=45.0, client_info=client_info, + self.cancel_operation, + default_timeout=45.0, + client_info=client_info, ), self.get_server_config: gapic_v1.method.wrap_method( self.get_server_config, @@ -247,7 +268,9 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.get_json_web_keys: gapic_v1.method.wrap_method( - self.get_json_web_keys, default_timeout=None, client_info=client_info, + self.get_json_web_keys, + default_timeout=None, + client_info=client_info, ), self.list_node_pools: gapic_v1.method.wrap_method( self.list_node_pools, @@ -280,7 +303,9 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.create_node_pool: gapic_v1.method.wrap_method( - self.create_node_pool, default_timeout=45.0, client_info=client_info, + self.create_node_pool, + default_timeout=45.0, + client_info=client_info, ), self.delete_node_pool: gapic_v1.method.wrap_method( self.delete_node_pool, @@ -308,13 +333,19 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.set_labels: gapic_v1.method.wrap_method( - self.set_labels, default_timeout=45.0, client_info=client_info, + self.set_labels, + default_timeout=45.0, + client_info=client_info, ), self.set_legacy_abac: gapic_v1.method.wrap_method( - self.set_legacy_abac, default_timeout=45.0, client_info=client_info, + self.set_legacy_abac, + default_timeout=45.0, + client_info=client_info, ), self.start_ip_rotation: gapic_v1.method.wrap_method( - self.start_ip_rotation, default_timeout=45.0, client_info=client_info, + self.start_ip_rotation, + default_timeout=45.0, + client_info=client_info, ), self.complete_ip_rotation: gapic_v1.method.wrap_method( self.complete_ip_rotation, @@ -322,10 +353,14 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.set_node_pool_size: gapic_v1.method.wrap_method( - self.set_node_pool_size, default_timeout=45.0, client_info=client_info, + self.set_node_pool_size, + default_timeout=45.0, + client_info=client_info, ), self.set_network_policy: gapic_v1.method.wrap_method( - self.set_network_policy, default_timeout=45.0, client_info=client_info, + self.set_network_policy, + default_timeout=45.0, + client_info=client_info, ), self.set_maintenance_policy: gapic_v1.method.wrap_method( self.set_maintenance_policy, @@ -342,9 +377,9 @@ def _prep_wrapped_messages(self, client_info): def close(self): """Closes resources associated with the transport. - .. warning:: - Only call this method if the transport is NOT shared - with other clients - this may cause errors in other clients! + .. warning:: + Only call this method if the transport is NOT shared + with other clients - this may cause errors in other clients! """ raise NotImplementedError() @@ -651,5 +686,9 @@ def list_usable_subnetworks( ]: raise NotImplementedError() + @property + def kind(self) -> str: + raise NotImplementedError() + __all__ = ("ClusterManagerTransport",) diff --git a/google/cloud/container_v1/services/cluster_manager/transports/grpc.py b/google/cloud/container_v1/services/cluster_manager/transports/grpc.py index 4986dae5..79643edc 100644 --- a/google/cloud/container_v1/services/cluster_manager/transports/grpc.py +++ b/google/cloud/container_v1/services/cluster_manager/transports/grpc.py @@ -13,20 +13,19 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import warnings from typing import Callable, Dict, Optional, Sequence, Tuple, Union +import warnings -from google.api_core import grpc_helpers -from google.api_core import gapic_v1 +from google.api_core import gapic_v1, grpc_helpers import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore - +from google.protobuf import empty_pb2 # type: ignore import grpc # type: ignore from google.cloud.container_v1.types import cluster_service -from google.protobuf import empty_pb2 # type: ignore -from .base import ClusterManagerTransport, DEFAULT_CLIENT_INFO + +from .base import DEFAULT_CLIENT_INFO, ClusterManagerTransport class ClusterManagerGrpcTransport(ClusterManagerTransport): @@ -225,8 +224,7 @@ def create_channel( @property def grpc_channel(self) -> grpc.Channel: - """Return the channel designed to connect to this service. - """ + """Return the channel designed to connect to this service.""" return self._grpc_channel @property @@ -1132,5 +1130,9 @@ def list_usable_subnetworks( def close(self): self.grpc_channel.close() + @property + def kind(self) -> str: + return "grpc" + __all__ = ("ClusterManagerGrpcTransport",) diff --git a/google/cloud/container_v1/services/cluster_manager/transports/grpc_asyncio.py b/google/cloud/container_v1/services/cluster_manager/transports/grpc_asyncio.py index 052036a7..a895fd76 100644 --- a/google/cloud/container_v1/services/cluster_manager/transports/grpc_asyncio.py +++ b/google/cloud/container_v1/services/cluster_manager/transports/grpc_asyncio.py @@ -13,20 +13,19 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import warnings from typing import Awaitable, Callable, Dict, Optional, Sequence, Tuple, Union +import warnings -from google.api_core import gapic_v1 -from google.api_core import grpc_helpers_async +from google.api_core import gapic_v1, grpc_helpers_async from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore - +from google.protobuf import empty_pb2 # type: ignore import grpc # type: ignore from grpc.experimental import aio # type: ignore from google.cloud.container_v1.types import cluster_service -from google.protobuf import empty_pb2 # type: ignore -from .base import ClusterManagerTransport, DEFAULT_CLIENT_INFO + +from .base import DEFAULT_CLIENT_INFO, ClusterManagerTransport from .grpc import ClusterManagerGrpcTransport diff --git a/google/cloud/container_v1/types/__init__.py b/google/cloud/container_v1/types/__init__.py index 89f6ba10..a6e86485 100644 --- a/google/cloud/container_v1/types/__init__.py +++ b/google/cloud/container_v1/types/__init__.py @@ -35,6 +35,7 @@ CreateNodePoolRequest, DailyMaintenanceWindow, DatabaseEncryption, + DatapathProvider, DefaultSnatStatus, DeleteClusterRequest, DeleteNodePoolRequest, @@ -96,6 +97,7 @@ OperationProgress, PrivateClusterConfig, PrivateClusterMasterGlobalAccessConfig, + PrivateIPv6GoogleAccess, RecurringTimeWindow, ReleaseChannel, ReservationAffinity, @@ -128,15 +130,13 @@ UpdateNodePoolRequest, UpgradeAvailableEvent, UpgradeEvent, + UpgradeResourceType, UsableSubnetwork, UsableSubnetworkSecondaryRange, VerticalPodAutoscaling, VirtualNIC, WorkloadIdentityConfig, WorkloadMetadataConfig, - DatapathProvider, - PrivateIPv6GoogleAccess, - UpgradeResourceType, ) __all__ = ( diff --git a/google/cloud/container_v1/types/cluster_service.py b/google/cloud/container_v1/types/cluster_service.py index 0a9e096a..33f1b470 100644 --- a/google/cloud/container_v1/types/cluster_service.py +++ b/google/cloud/container_v1/types/cluster_service.py @@ -13,13 +13,11 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import proto # type: ignore - from google.protobuf import timestamp_pb2 # type: ignore from google.protobuf import wrappers_pb2 # type: ignore from google.rpc import code_pb2 # type: ignore from google.rpc import status_pb2 # type: ignore - +import proto # type: ignore __protobuf__ = proto.module( package="google.container.v1", @@ -184,7 +182,7 @@ class LinuxNodeConfig(proto.Message): r"""Parameters that can be configured on Linux nodes. Attributes: - sysctls (Sequence[google.cloud.container_v1.types.LinuxNodeConfig.SysctlsEntry]): + sysctls (Mapping[str, str]): The Linux kernel parameters to be applied to the nodes and all pods running on the nodes. @@ -196,7 +194,11 @@ class LinuxNodeConfig(proto.Message): net.ipv4.tcp_tw_reuse """ - sysctls = proto.MapField(proto.STRING, proto.STRING, number=1,) + sysctls = proto.MapField( + proto.STRING, + proto.STRING, + number=1, + ) class NodeKubeletConfig(proto.Message): @@ -238,11 +240,19 @@ class NodeKubeletConfig(proto.Message): The value must be a positive duration. """ - cpu_manager_policy = proto.Field(proto.STRING, number=1,) + cpu_manager_policy = proto.Field( + proto.STRING, + number=1, + ) cpu_cfs_quota = proto.Field( - proto.MESSAGE, number=2, message=wrappers_pb2.BoolValue, + proto.MESSAGE, + number=2, + message=wrappers_pb2.BoolValue, + ) + cpu_cfs_quota_period = proto.Field( + proto.STRING, + number=3, ) - cpu_cfs_quota_period = proto.Field(proto.STRING, number=3,) class NodeConfig(proto.Message): @@ -282,7 +292,7 @@ class NodeConfig(proto.Message): address of the Service Account; otherwise, if no Service Account is specified, the "default" service account is used. - metadata (Sequence[google.cloud.container_v1.types.NodeConfig.MetadataEntry]): + metadata (Mapping[str, str]): The metadata key/value pairs assigned to instances in the cluster. @@ -323,7 +333,7 @@ class NodeConfig(proto.Message): The image type to use for this node. Note that for a given image type, the latest version of it will be used. - labels (Sequence[google.cloud.container_v1.types.NodeConfig.LabelsEntry]): + labels (Mapping[str, str]): The map of Kubernetes labels (key/value pairs) to be applied to each node. These will added in addition to any default label(s) that @@ -420,43 +430,119 @@ class NodeConfig(proto.Message): Enable or disable gvnic in the node pool. """ - machine_type = proto.Field(proto.STRING, number=1,) - disk_size_gb = proto.Field(proto.INT32, number=2,) - oauth_scopes = proto.RepeatedField(proto.STRING, number=3,) - service_account = proto.Field(proto.STRING, number=9,) - metadata = proto.MapField(proto.STRING, proto.STRING, number=4,) - image_type = proto.Field(proto.STRING, number=5,) - labels = proto.MapField(proto.STRING, proto.STRING, number=6,) - local_ssd_count = proto.Field(proto.INT32, number=7,) - tags = proto.RepeatedField(proto.STRING, number=8,) - preemptible = proto.Field(proto.BOOL, number=10,) + machine_type = proto.Field( + proto.STRING, + number=1, + ) + disk_size_gb = proto.Field( + proto.INT32, + number=2, + ) + oauth_scopes = proto.RepeatedField( + proto.STRING, + number=3, + ) + service_account = proto.Field( + proto.STRING, + number=9, + ) + metadata = proto.MapField( + proto.STRING, + proto.STRING, + number=4, + ) + image_type = proto.Field( + proto.STRING, + number=5, + ) + labels = proto.MapField( + proto.STRING, + proto.STRING, + number=6, + ) + local_ssd_count = proto.Field( + proto.INT32, + number=7, + ) + tags = proto.RepeatedField( + proto.STRING, + number=8, + ) + preemptible = proto.Field( + proto.BOOL, + number=10, + ) accelerators = proto.RepeatedField( - proto.MESSAGE, number=11, message="AcceleratorConfig", + proto.MESSAGE, + number=11, + message="AcceleratorConfig", + ) + disk_type = proto.Field( + proto.STRING, + number=12, + ) + min_cpu_platform = proto.Field( + proto.STRING, + number=13, ) - disk_type = proto.Field(proto.STRING, number=12,) - min_cpu_platform = proto.Field(proto.STRING, number=13,) workload_metadata_config = proto.Field( - proto.MESSAGE, number=14, message="WorkloadMetadataConfig", + proto.MESSAGE, + number=14, + message="WorkloadMetadataConfig", + ) + taints = proto.RepeatedField( + proto.MESSAGE, + number=15, + message="NodeTaint", + ) + sandbox_config = proto.Field( + proto.MESSAGE, + number=17, + message="SandboxConfig", + ) + node_group = proto.Field( + proto.STRING, + number=18, ) - taints = proto.RepeatedField(proto.MESSAGE, number=15, message="NodeTaint",) - sandbox_config = proto.Field(proto.MESSAGE, number=17, message="SandboxConfig",) - node_group = proto.Field(proto.STRING, number=18,) reservation_affinity = proto.Field( - proto.MESSAGE, number=19, message="ReservationAffinity", + proto.MESSAGE, + number=19, + message="ReservationAffinity", ) shielded_instance_config = proto.Field( - proto.MESSAGE, number=20, message="ShieldedInstanceConfig", + proto.MESSAGE, + number=20, + message="ShieldedInstanceConfig", ) linux_node_config = proto.Field( - proto.MESSAGE, number=21, message="LinuxNodeConfig", + proto.MESSAGE, + number=21, + message="LinuxNodeConfig", + ) + kubelet_config = proto.Field( + proto.MESSAGE, + number=22, + message="NodeKubeletConfig", + ) + boot_disk_kms_key = proto.Field( + proto.STRING, + number=23, + ) + gcfs_config = proto.Field( + proto.MESSAGE, + number=25, + message="GcfsConfig", ) - kubelet_config = proto.Field(proto.MESSAGE, number=22, message="NodeKubeletConfig",) - boot_disk_kms_key = proto.Field(proto.STRING, number=23,) - gcfs_config = proto.Field(proto.MESSAGE, number=25, message="GcfsConfig",) advanced_machine_features = proto.Field( - proto.MESSAGE, number=26, message="AdvancedMachineFeatures", + proto.MESSAGE, + number=26, + message="AdvancedMachineFeatures", + ) + gvnic = proto.Field( + proto.MESSAGE, + number=29, + message="VirtualNIC", ) - gvnic = proto.Field(proto.MESSAGE, number=29, message="VirtualNIC",) class AdvancedMachineFeatures(proto.Message): @@ -473,7 +559,11 @@ class AdvancedMachineFeatures(proto.Message): This field is a member of `oneof`_ ``_threads_per_core``. """ - threads_per_core = proto.Field(proto.INT64, number=1, optional=True,) + threads_per_core = proto.Field( + proto.INT64, + number=1, + optional=True, + ) class NodeNetworkConfig(proto.Message): @@ -527,9 +617,18 @@ class NodeNetworkConfig(proto.Message): created. """ - create_pod_range = proto.Field(proto.BOOL, number=4,) - pod_range = proto.Field(proto.STRING, number=5,) - pod_ipv4_cidr_block = proto.Field(proto.STRING, number=6,) + create_pod_range = proto.Field( + proto.BOOL, + number=4, + ) + pod_range = proto.Field( + proto.STRING, + number=5, + ) + pod_ipv4_cidr_block = proto.Field( + proto.STRING, + number=6, + ) class ShieldedInstanceConfig(proto.Message): @@ -555,8 +654,14 @@ class ShieldedInstanceConfig(proto.Message): is created. """ - enable_secure_boot = proto.Field(proto.BOOL, number=1,) - enable_integrity_monitoring = proto.Field(proto.BOOL, number=2,) + enable_secure_boot = proto.Field( + proto.BOOL, + number=1, + ) + enable_integrity_monitoring = proto.Field( + proto.BOOL, + number=2, + ) class SandboxConfig(proto.Message): @@ -573,7 +678,11 @@ class Type(proto.Enum): UNSPECIFIED = 0 GVISOR = 1 - type_ = proto.Field(proto.ENUM, number=2, enum=Type,) + type_ = proto.Field( + proto.ENUM, + number=2, + enum=Type, + ) class GcfsConfig(proto.Message): @@ -585,7 +694,10 @@ class GcfsConfig(proto.Message): Whether to use GCFS. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class ReservationAffinity(proto.Message): @@ -616,9 +728,19 @@ class Type(proto.Enum): ANY_RESERVATION = 2 SPECIFIC_RESERVATION = 3 - consume_reservation_type = proto.Field(proto.ENUM, number=1, enum=Type,) - key = proto.Field(proto.STRING, number=2,) - values = proto.RepeatedField(proto.STRING, number=3,) + consume_reservation_type = proto.Field( + proto.ENUM, + number=1, + enum=Type, + ) + key = proto.Field( + proto.STRING, + number=2, + ) + values = proto.RepeatedField( + proto.STRING, + number=3, + ) class NodeTaint(proto.Message): @@ -646,9 +768,19 @@ class Effect(proto.Enum): PREFER_NO_SCHEDULE = 2 NO_EXECUTE = 3 - key = proto.Field(proto.STRING, number=1,) - value = proto.Field(proto.STRING, number=2,) - effect = proto.Field(proto.ENUM, number=3, enum=Effect,) + key = proto.Field( + proto.STRING, + number=1, + ) + value = proto.Field( + proto.STRING, + number=2, + ) + effect = proto.Field( + proto.ENUM, + number=3, + enum=Effect, + ) class MasterAuth(proto.Message): @@ -697,14 +829,31 @@ class MasterAuth(proto.Message): authenticate to the cluster endpoint. """ - username = proto.Field(proto.STRING, number=1,) - password = proto.Field(proto.STRING, number=2,) + username = proto.Field( + proto.STRING, + number=1, + ) + password = proto.Field( + proto.STRING, + number=2, + ) client_certificate_config = proto.Field( - proto.MESSAGE, number=3, message="ClientCertificateConfig", + proto.MESSAGE, + number=3, + message="ClientCertificateConfig", + ) + cluster_ca_certificate = proto.Field( + proto.STRING, + number=100, + ) + client_certificate = proto.Field( + proto.STRING, + number=101, + ) + client_key = proto.Field( + proto.STRING, + number=102, ) - cluster_ca_certificate = proto.Field(proto.STRING, number=100,) - client_certificate = proto.Field(proto.STRING, number=101,) - client_key = proto.Field(proto.STRING, number=102,) class ClientCertificateConfig(proto.Message): @@ -715,7 +864,10 @@ class ClientCertificateConfig(proto.Message): Issue a client certificate. """ - issue_client_certificate = proto.Field(proto.BOOL, number=1,) + issue_client_certificate = proto.Field( + proto.BOOL, + number=1, + ) class AddonsConfig(proto.Message): @@ -767,27 +919,49 @@ class AddonsConfig(proto.Message): """ http_load_balancing = proto.Field( - proto.MESSAGE, number=1, message="HttpLoadBalancing", + proto.MESSAGE, + number=1, + message="HttpLoadBalancing", ) horizontal_pod_autoscaling = proto.Field( - proto.MESSAGE, number=2, message="HorizontalPodAutoscaling", + proto.MESSAGE, + number=2, + message="HorizontalPodAutoscaling", ) kubernetes_dashboard = proto.Field( - proto.MESSAGE, number=3, message="KubernetesDashboard", + proto.MESSAGE, + number=3, + message="KubernetesDashboard", ) network_policy_config = proto.Field( - proto.MESSAGE, number=4, message="NetworkPolicyConfig", + proto.MESSAGE, + number=4, + message="NetworkPolicyConfig", + ) + cloud_run_config = proto.Field( + proto.MESSAGE, + number=7, + message="CloudRunConfig", + ) + dns_cache_config = proto.Field( + proto.MESSAGE, + number=8, + message="DnsCacheConfig", ) - cloud_run_config = proto.Field(proto.MESSAGE, number=7, message="CloudRunConfig",) - dns_cache_config = proto.Field(proto.MESSAGE, number=8, message="DnsCacheConfig",) config_connector_config = proto.Field( - proto.MESSAGE, number=10, message="ConfigConnectorConfig", + proto.MESSAGE, + number=10, + message="ConfigConnectorConfig", ) gce_persistent_disk_csi_driver_config = proto.Field( - proto.MESSAGE, number=11, message="GcePersistentDiskCsiDriverConfig", + proto.MESSAGE, + number=11, + message="GcePersistentDiskCsiDriverConfig", ) gcp_filestore_csi_driver_config = proto.Field( - proto.MESSAGE, number=14, message="GcpFilestoreCsiDriverConfig", + proto.MESSAGE, + number=14, + message="GcpFilestoreCsiDriverConfig", ) @@ -804,7 +978,10 @@ class HttpLoadBalancing(proto.Message): balancers. """ - disabled = proto.Field(proto.BOOL, number=1,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) class HorizontalPodAutoscaling(proto.Message): @@ -821,7 +998,10 @@ class HorizontalPodAutoscaling(proto.Message): Stackdriver Monitoring. """ - disabled = proto.Field(proto.BOOL, number=1,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) class KubernetesDashboard(proto.Message): @@ -833,7 +1013,10 @@ class KubernetesDashboard(proto.Message): for this cluster. """ - disabled = proto.Field(proto.BOOL, number=1,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) class NetworkPolicyConfig(proto.Message): @@ -847,7 +1030,10 @@ class NetworkPolicyConfig(proto.Message): cluster. """ - disabled = proto.Field(proto.BOOL, number=1,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) class DnsCacheConfig(proto.Message): @@ -859,7 +1045,10 @@ class DnsCacheConfig(proto.Message): this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class PrivateClusterMasterGlobalAccessConfig(proto.Message): @@ -871,7 +1060,10 @@ class PrivateClusterMasterGlobalAccessConfig(proto.Message): not. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class PrivateClusterConfig(proto.Message): @@ -906,14 +1098,34 @@ class PrivateClusterConfig(proto.Message): Controls master global access settings. """ - enable_private_nodes = proto.Field(proto.BOOL, number=1,) - enable_private_endpoint = proto.Field(proto.BOOL, number=2,) - master_ipv4_cidr_block = proto.Field(proto.STRING, number=3,) - private_endpoint = proto.Field(proto.STRING, number=4,) - public_endpoint = proto.Field(proto.STRING, number=5,) - peering_name = proto.Field(proto.STRING, number=7,) + enable_private_nodes = proto.Field( + proto.BOOL, + number=1, + ) + enable_private_endpoint = proto.Field( + proto.BOOL, + number=2, + ) + master_ipv4_cidr_block = proto.Field( + proto.STRING, + number=3, + ) + private_endpoint = proto.Field( + proto.STRING, + number=4, + ) + public_endpoint = proto.Field( + proto.STRING, + number=5, + ) + peering_name = proto.Field( + proto.STRING, + number=7, + ) master_global_access_config = proto.Field( - proto.MESSAGE, number=8, message="PrivateClusterMasterGlobalAccessConfig", + proto.MESSAGE, + number=8, + message="PrivateClusterMasterGlobalAccessConfig", ) @@ -931,8 +1143,14 @@ class AuthenticatorGroupsConfig(proto.Message): be used. Only relevant if enabled = true. """ - enabled = proto.Field(proto.BOOL, number=1,) - security_group = proto.Field(proto.STRING, number=2,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) + security_group = proto.Field( + proto.STRING, + number=2, + ) class CloudRunConfig(proto.Message): @@ -953,8 +1171,15 @@ class LoadBalancerType(proto.Enum): LOAD_BALANCER_TYPE_EXTERNAL = 1 LOAD_BALANCER_TYPE_INTERNAL = 2 - disabled = proto.Field(proto.BOOL, number=1,) - load_balancer_type = proto.Field(proto.ENUM, number=3, enum=LoadBalancerType,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) + load_balancer_type = proto.Field( + proto.ENUM, + number=3, + enum=LoadBalancerType, + ) class ConfigConnectorConfig(proto.Message): @@ -966,7 +1191,10 @@ class ConfigConnectorConfig(proto.Message): cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class GcePersistentDiskCsiDriverConfig(proto.Message): @@ -978,7 +1206,10 @@ class GcePersistentDiskCsiDriverConfig(proto.Message): enabled for this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class GcpFilestoreCsiDriverConfig(proto.Message): @@ -990,7 +1221,10 @@ class GcpFilestoreCsiDriverConfig(proto.Message): enabled for this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class MasterAuthorizedNetworksConfig(proto.Message): @@ -1020,11 +1254,24 @@ class CidrBlock(proto.Message): cidr_block must be specified in CIDR notation. """ - display_name = proto.Field(proto.STRING, number=1,) - cidr_block = proto.Field(proto.STRING, number=2,) + display_name = proto.Field( + proto.STRING, + number=1, + ) + cidr_block = proto.Field( + proto.STRING, + number=2, + ) - enabled = proto.Field(proto.BOOL, number=1,) - cidr_blocks = proto.RepeatedField(proto.MESSAGE, number=2, message=CidrBlock,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) + cidr_blocks = proto.RepeatedField( + proto.MESSAGE, + number=2, + message=CidrBlock, + ) class LegacyAbac(proto.Message): @@ -1041,7 +1288,10 @@ class LegacyAbac(proto.Message): configuration or IAM. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class NetworkPolicy(proto.Message): @@ -1061,8 +1311,15 @@ class Provider(proto.Enum): PROVIDER_UNSPECIFIED = 0 CALICO = 1 - provider = proto.Field(proto.ENUM, number=1, enum=Provider,) - enabled = proto.Field(proto.BOOL, number=2,) + provider = proto.Field( + proto.ENUM, + number=1, + enum=Provider, + ) + enabled = proto.Field( + proto.BOOL, + number=2, + ) class BinaryAuthorization(proto.Message): @@ -1075,7 +1332,10 @@ class BinaryAuthorization(proto.Message): validated by Binary Authorization. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class IPAllocationPolicy(proto.Message): @@ -1198,19 +1458,58 @@ class IPAllocationPolicy(proto.Message): allocation mode """ - use_ip_aliases = proto.Field(proto.BOOL, number=1,) - create_subnetwork = proto.Field(proto.BOOL, number=2,) - subnetwork_name = proto.Field(proto.STRING, number=3,) - cluster_ipv4_cidr = proto.Field(proto.STRING, number=4,) - node_ipv4_cidr = proto.Field(proto.STRING, number=5,) - services_ipv4_cidr = proto.Field(proto.STRING, number=6,) - cluster_secondary_range_name = proto.Field(proto.STRING, number=7,) - services_secondary_range_name = proto.Field(proto.STRING, number=8,) - cluster_ipv4_cidr_block = proto.Field(proto.STRING, number=9,) - node_ipv4_cidr_block = proto.Field(proto.STRING, number=10,) - services_ipv4_cidr_block = proto.Field(proto.STRING, number=11,) - tpu_ipv4_cidr_block = proto.Field(proto.STRING, number=13,) - use_routes = proto.Field(proto.BOOL, number=15,) + use_ip_aliases = proto.Field( + proto.BOOL, + number=1, + ) + create_subnetwork = proto.Field( + proto.BOOL, + number=2, + ) + subnetwork_name = proto.Field( + proto.STRING, + number=3, + ) + cluster_ipv4_cidr = proto.Field( + proto.STRING, + number=4, + ) + node_ipv4_cidr = proto.Field( + proto.STRING, + number=5, + ) + services_ipv4_cidr = proto.Field( + proto.STRING, + number=6, + ) + cluster_secondary_range_name = proto.Field( + proto.STRING, + number=7, + ) + services_secondary_range_name = proto.Field( + proto.STRING, + number=8, + ) + cluster_ipv4_cidr_block = proto.Field( + proto.STRING, + number=9, + ) + node_ipv4_cidr_block = proto.Field( + proto.STRING, + number=10, + ) + services_ipv4_cidr_block = proto.Field( + proto.STRING, + number=11, + ) + tpu_ipv4_cidr_block = proto.Field( + proto.STRING, + number=13, + ) + use_routes = proto.Field( + proto.BOOL, + number=15, + ) class Cluster(proto.Message): @@ -1333,7 +1632,7 @@ class Cluster(proto.Message): uptime and master/node upgrades are disabled. Alpha enabled clusters are automatically deleted thirty days after creation. - resource_labels (Sequence[google.cloud.container_v1.types.Cluster.ResourceLabelsEntry]): + resource_labels (Mapping[str, str]): The resource labels for the cluster to use to annotate any related Google Compute Engine resources. @@ -1510,98 +1809,273 @@ class Status(proto.Enum): ERROR = 5 DEGRADED = 6 - name = proto.Field(proto.STRING, number=1,) - description = proto.Field(proto.STRING, number=2,) - initial_node_count = proto.Field(proto.INT32, number=3,) - node_config = proto.Field(proto.MESSAGE, number=4, message="NodeConfig",) - master_auth = proto.Field(proto.MESSAGE, number=5, message="MasterAuth",) - logging_service = proto.Field(proto.STRING, number=6,) - monitoring_service = proto.Field(proto.STRING, number=7,) - network = proto.Field(proto.STRING, number=8,) - cluster_ipv4_cidr = proto.Field(proto.STRING, number=9,) - addons_config = proto.Field(proto.MESSAGE, number=10, message="AddonsConfig",) - subnetwork = proto.Field(proto.STRING, number=11,) - node_pools = proto.RepeatedField(proto.MESSAGE, number=12, message="NodePool",) - locations = proto.RepeatedField(proto.STRING, number=13,) - enable_kubernetes_alpha = proto.Field(proto.BOOL, number=14,) - resource_labels = proto.MapField(proto.STRING, proto.STRING, number=15,) - label_fingerprint = proto.Field(proto.STRING, number=16,) - legacy_abac = proto.Field(proto.MESSAGE, number=18, message="LegacyAbac",) - network_policy = proto.Field(proto.MESSAGE, number=19, message="NetworkPolicy",) + name = proto.Field( + proto.STRING, + number=1, + ) + description = proto.Field( + proto.STRING, + number=2, + ) + initial_node_count = proto.Field( + proto.INT32, + number=3, + ) + node_config = proto.Field( + proto.MESSAGE, + number=4, + message="NodeConfig", + ) + master_auth = proto.Field( + proto.MESSAGE, + number=5, + message="MasterAuth", + ) + logging_service = proto.Field( + proto.STRING, + number=6, + ) + monitoring_service = proto.Field( + proto.STRING, + number=7, + ) + network = proto.Field( + proto.STRING, + number=8, + ) + cluster_ipv4_cidr = proto.Field( + proto.STRING, + number=9, + ) + addons_config = proto.Field( + proto.MESSAGE, + number=10, + message="AddonsConfig", + ) + subnetwork = proto.Field( + proto.STRING, + number=11, + ) + node_pools = proto.RepeatedField( + proto.MESSAGE, + number=12, + message="NodePool", + ) + locations = proto.RepeatedField( + proto.STRING, + number=13, + ) + enable_kubernetes_alpha = proto.Field( + proto.BOOL, + number=14, + ) + resource_labels = proto.MapField( + proto.STRING, + proto.STRING, + number=15, + ) + label_fingerprint = proto.Field( + proto.STRING, + number=16, + ) + legacy_abac = proto.Field( + proto.MESSAGE, + number=18, + message="LegacyAbac", + ) + network_policy = proto.Field( + proto.MESSAGE, + number=19, + message="NetworkPolicy", + ) ip_allocation_policy = proto.Field( - proto.MESSAGE, number=20, message="IPAllocationPolicy", + proto.MESSAGE, + number=20, + message="IPAllocationPolicy", ) master_authorized_networks_config = proto.Field( - proto.MESSAGE, number=22, message="MasterAuthorizedNetworksConfig", + proto.MESSAGE, + number=22, + message="MasterAuthorizedNetworksConfig", ) maintenance_policy = proto.Field( - proto.MESSAGE, number=23, message="MaintenancePolicy", + proto.MESSAGE, + number=23, + message="MaintenancePolicy", ) binary_authorization = proto.Field( - proto.MESSAGE, number=24, message="BinaryAuthorization", + proto.MESSAGE, + number=24, + message="BinaryAuthorization", + ) + autoscaling = proto.Field( + proto.MESSAGE, + number=26, + message="ClusterAutoscaling", + ) + network_config = proto.Field( + proto.MESSAGE, + number=27, + message="NetworkConfig", ) - autoscaling = proto.Field(proto.MESSAGE, number=26, message="ClusterAutoscaling",) - network_config = proto.Field(proto.MESSAGE, number=27, message="NetworkConfig",) default_max_pods_constraint = proto.Field( - proto.MESSAGE, number=30, message="MaxPodsConstraint", + proto.MESSAGE, + number=30, + message="MaxPodsConstraint", ) resource_usage_export_config = proto.Field( - proto.MESSAGE, number=33, message="ResourceUsageExportConfig", + proto.MESSAGE, + number=33, + message="ResourceUsageExportConfig", ) authenticator_groups_config = proto.Field( - proto.MESSAGE, number=34, message="AuthenticatorGroupsConfig", + proto.MESSAGE, + number=34, + message="AuthenticatorGroupsConfig", ) private_cluster_config = proto.Field( - proto.MESSAGE, number=37, message="PrivateClusterConfig", + proto.MESSAGE, + number=37, + message="PrivateClusterConfig", ) database_encryption = proto.Field( - proto.MESSAGE, number=38, message="DatabaseEncryption", + proto.MESSAGE, + number=38, + message="DatabaseEncryption", ) vertical_pod_autoscaling = proto.Field( - proto.MESSAGE, number=39, message="VerticalPodAutoscaling", + proto.MESSAGE, + number=39, + message="VerticalPodAutoscaling", + ) + shielded_nodes = proto.Field( + proto.MESSAGE, + number=40, + message="ShieldedNodes", + ) + release_channel = proto.Field( + proto.MESSAGE, + number=41, + message="ReleaseChannel", ) - shielded_nodes = proto.Field(proto.MESSAGE, number=40, message="ShieldedNodes",) - release_channel = proto.Field(proto.MESSAGE, number=41, message="ReleaseChannel",) workload_identity_config = proto.Field( - proto.MESSAGE, number=43, message="WorkloadIdentityConfig", + proto.MESSAGE, + number=43, + message="WorkloadIdentityConfig", ) mesh_certificates = proto.Field( - proto.MESSAGE, number=67, message="MeshCertificates", + proto.MESSAGE, + number=67, + message="MeshCertificates", ) notification_config = proto.Field( - proto.MESSAGE, number=49, message="NotificationConfig", + proto.MESSAGE, + number=49, + message="NotificationConfig", ) confidential_nodes = proto.Field( - proto.MESSAGE, number=50, message="ConfidentialNodes", - ) - self_link = proto.Field(proto.STRING, number=100,) - zone = proto.Field(proto.STRING, number=101,) - endpoint = proto.Field(proto.STRING, number=102,) - initial_cluster_version = proto.Field(proto.STRING, number=103,) - current_master_version = proto.Field(proto.STRING, number=104,) - current_node_version = proto.Field(proto.STRING, number=105,) - create_time = proto.Field(proto.STRING, number=106,) - status = proto.Field(proto.ENUM, number=107, enum=Status,) - status_message = proto.Field(proto.STRING, number=108,) - node_ipv4_cidr_size = proto.Field(proto.INT32, number=109,) - services_ipv4_cidr = proto.Field(proto.STRING, number=110,) - instance_group_urls = proto.RepeatedField(proto.STRING, number=111,) - current_node_count = proto.Field(proto.INT32, number=112,) - expire_time = proto.Field(proto.STRING, number=113,) - location = proto.Field(proto.STRING, number=114,) - enable_tpu = proto.Field(proto.BOOL, number=115,) - tpu_ipv4_cidr_block = proto.Field(proto.STRING, number=116,) + proto.MESSAGE, + number=50, + message="ConfidentialNodes", + ) + self_link = proto.Field( + proto.STRING, + number=100, + ) + zone = proto.Field( + proto.STRING, + number=101, + ) + endpoint = proto.Field( + proto.STRING, + number=102, + ) + initial_cluster_version = proto.Field( + proto.STRING, + number=103, + ) + current_master_version = proto.Field( + proto.STRING, + number=104, + ) + current_node_version = proto.Field( + proto.STRING, + number=105, + ) + create_time = proto.Field( + proto.STRING, + number=106, + ) + status = proto.Field( + proto.ENUM, + number=107, + enum=Status, + ) + status_message = proto.Field( + proto.STRING, + number=108, + ) + node_ipv4_cidr_size = proto.Field( + proto.INT32, + number=109, + ) + services_ipv4_cidr = proto.Field( + proto.STRING, + number=110, + ) + instance_group_urls = proto.RepeatedField( + proto.STRING, + number=111, + ) + current_node_count = proto.Field( + proto.INT32, + number=112, + ) + expire_time = proto.Field( + proto.STRING, + number=113, + ) + location = proto.Field( + proto.STRING, + number=114, + ) + enable_tpu = proto.Field( + proto.BOOL, + number=115, + ) + tpu_ipv4_cidr_block = proto.Field( + proto.STRING, + number=116, + ) conditions = proto.RepeatedField( - proto.MESSAGE, number=118, message="StatusCondition", + proto.MESSAGE, + number=118, + message="StatusCondition", + ) + autopilot = proto.Field( + proto.MESSAGE, + number=128, + message="Autopilot", + ) + id = proto.Field( + proto.STRING, + number=129, ) - autopilot = proto.Field(proto.MESSAGE, number=128, message="Autopilot",) - id = proto.Field(proto.STRING, number=129,) node_pool_defaults = proto.Field( - proto.MESSAGE, number=131, optional=True, message="NodePoolDefaults", + proto.MESSAGE, + number=131, + optional=True, + message="NodePoolDefaults", + ) + logging_config = proto.Field( + proto.MESSAGE, + number=132, + message="LoggingConfig", ) - logging_config = proto.Field(proto.MESSAGE, number=132, message="LoggingConfig",) monitoring_config = proto.Field( - proto.MESSAGE, number=133, message="MonitoringConfig", + proto.MESSAGE, + number=133, + message="MonitoringConfig", ) @@ -1615,7 +2089,9 @@ class NodePoolDefaults(proto.Message): """ node_config_defaults = proto.Field( - proto.MESSAGE, number=1, message="NodeConfigDefaults", + proto.MESSAGE, + number=1, + message="NodeConfigDefaults", ) @@ -1628,7 +2104,11 @@ class NodeConfigDefaults(proto.Message): Riptide) options. """ - gcfs_config = proto.Field(proto.MESSAGE, number=1, message="GcfsConfig",) + gcfs_config = proto.Field( + proto.MESSAGE, + number=1, + message="GcfsConfig", + ) class ClusterUpdate(proto.Message): @@ -1779,84 +2259,159 @@ class ClusterUpdate(proto.Message): The desired GCFS config for the cluster """ - desired_node_version = proto.Field(proto.STRING, number=4,) - desired_monitoring_service = proto.Field(proto.STRING, number=5,) + desired_node_version = proto.Field( + proto.STRING, + number=4, + ) + desired_monitoring_service = proto.Field( + proto.STRING, + number=5, + ) desired_addons_config = proto.Field( - proto.MESSAGE, number=6, message="AddonsConfig", + proto.MESSAGE, + number=6, + message="AddonsConfig", + ) + desired_node_pool_id = proto.Field( + proto.STRING, + number=7, + ) + desired_image_type = proto.Field( + proto.STRING, + number=8, ) - desired_node_pool_id = proto.Field(proto.STRING, number=7,) - desired_image_type = proto.Field(proto.STRING, number=8,) desired_database_encryption = proto.Field( - proto.MESSAGE, number=46, message="DatabaseEncryption", + proto.MESSAGE, + number=46, + message="DatabaseEncryption", ) desired_workload_identity_config = proto.Field( - proto.MESSAGE, number=47, message="WorkloadIdentityConfig", + proto.MESSAGE, + number=47, + message="WorkloadIdentityConfig", ) desired_mesh_certificates = proto.Field( - proto.MESSAGE, number=67, message="MeshCertificates", + proto.MESSAGE, + number=67, + message="MeshCertificates", ) desired_shielded_nodes = proto.Field( - proto.MESSAGE, number=48, message="ShieldedNodes", + proto.MESSAGE, + number=48, + message="ShieldedNodes", + ) + desired_dns_config = proto.Field( + proto.MESSAGE, + number=53, + message="DNSConfig", ) - desired_dns_config = proto.Field(proto.MESSAGE, number=53, message="DNSConfig",) desired_node_pool_autoscaling = proto.Field( - proto.MESSAGE, number=9, message="NodePoolAutoscaling", + proto.MESSAGE, + number=9, + message="NodePoolAutoscaling", + ) + desired_locations = proto.RepeatedField( + proto.STRING, + number=10, ) - desired_locations = proto.RepeatedField(proto.STRING, number=10,) desired_master_authorized_networks_config = proto.Field( - proto.MESSAGE, number=12, message="MasterAuthorizedNetworksConfig", + proto.MESSAGE, + number=12, + message="MasterAuthorizedNetworksConfig", ) desired_cluster_autoscaling = proto.Field( - proto.MESSAGE, number=15, message="ClusterAutoscaling", + proto.MESSAGE, + number=15, + message="ClusterAutoscaling", ) desired_binary_authorization = proto.Field( - proto.MESSAGE, number=16, message="BinaryAuthorization", + proto.MESSAGE, + number=16, + message="BinaryAuthorization", + ) + desired_logging_service = proto.Field( + proto.STRING, + number=19, ) - desired_logging_service = proto.Field(proto.STRING, number=19,) desired_resource_usage_export_config = proto.Field( - proto.MESSAGE, number=21, message="ResourceUsageExportConfig", + proto.MESSAGE, + number=21, + message="ResourceUsageExportConfig", ) desired_vertical_pod_autoscaling = proto.Field( - proto.MESSAGE, number=22, message="VerticalPodAutoscaling", + proto.MESSAGE, + number=22, + message="VerticalPodAutoscaling", ) desired_private_cluster_config = proto.Field( - proto.MESSAGE, number=25, message="PrivateClusterConfig", + proto.MESSAGE, + number=25, + message="PrivateClusterConfig", ) desired_intra_node_visibility_config = proto.Field( - proto.MESSAGE, number=26, message="IntraNodeVisibilityConfig", + proto.MESSAGE, + number=26, + message="IntraNodeVisibilityConfig", ) desired_default_snat_status = proto.Field( - proto.MESSAGE, number=28, message="DefaultSnatStatus", + proto.MESSAGE, + number=28, + message="DefaultSnatStatus", ) desired_release_channel = proto.Field( - proto.MESSAGE, number=31, message="ReleaseChannel", + proto.MESSAGE, + number=31, + message="ReleaseChannel", ) desired_l4ilb_subsetting_config = proto.Field( - proto.MESSAGE, number=39, message="ILBSubsettingConfig", + proto.MESSAGE, + number=39, + message="ILBSubsettingConfig", ) desired_datapath_provider = proto.Field( - proto.ENUM, number=50, enum="DatapathProvider", + proto.ENUM, + number=50, + enum="DatapathProvider", ) desired_private_ipv6_google_access = proto.Field( - proto.ENUM, number=51, enum="PrivateIPv6GoogleAccess", + proto.ENUM, + number=51, + enum="PrivateIPv6GoogleAccess", ) desired_notification_config = proto.Field( - proto.MESSAGE, number=55, message="NotificationConfig", + proto.MESSAGE, + number=55, + message="NotificationConfig", ) desired_authenticator_groups_config = proto.Field( - proto.MESSAGE, number=63, message="AuthenticatorGroupsConfig", + proto.MESSAGE, + number=63, + message="AuthenticatorGroupsConfig", ) desired_logging_config = proto.Field( - proto.MESSAGE, number=64, message="LoggingConfig", + proto.MESSAGE, + number=64, + message="LoggingConfig", ) desired_monitoring_config = proto.Field( - proto.MESSAGE, number=65, message="MonitoringConfig", + proto.MESSAGE, + number=65, + message="MonitoringConfig", ) desired_service_external_ips_config = proto.Field( - proto.MESSAGE, number=60, message="ServiceExternalIPsConfig", + proto.MESSAGE, + number=60, + message="ServiceExternalIPsConfig", + ) + desired_master_version = proto.Field( + proto.STRING, + number=100, + ) + desired_gcfs_config = proto.Field( + proto.MESSAGE, + number=109, + message="GcfsConfig", ) - desired_master_version = proto.Field(proto.STRING, number=100,) - desired_gcfs_config = proto.Field(proto.MESSAGE, number=109, message="GcfsConfig",) class Operation(proto.Message): @@ -1943,25 +2498,72 @@ class Type(proto.Enum): SET_NETWORK_POLICY = 15 SET_MAINTENANCE_POLICY = 16 - name = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - operation_type = proto.Field(proto.ENUM, number=3, enum=Type,) - status = proto.Field(proto.ENUM, number=4, enum=Status,) - detail = proto.Field(proto.STRING, number=8,) - status_message = proto.Field(proto.STRING, number=5,) - self_link = proto.Field(proto.STRING, number=6,) - target_link = proto.Field(proto.STRING, number=7,) - location = proto.Field(proto.STRING, number=9,) - start_time = proto.Field(proto.STRING, number=10,) - end_time = proto.Field(proto.STRING, number=11,) - progress = proto.Field(proto.MESSAGE, number=12, message="OperationProgress",) + name = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + operation_type = proto.Field( + proto.ENUM, + number=3, + enum=Type, + ) + status = proto.Field( + proto.ENUM, + number=4, + enum=Status, + ) + detail = proto.Field( + proto.STRING, + number=8, + ) + status_message = proto.Field( + proto.STRING, + number=5, + ) + self_link = proto.Field( + proto.STRING, + number=6, + ) + target_link = proto.Field( + proto.STRING, + number=7, + ) + location = proto.Field( + proto.STRING, + number=9, + ) + start_time = proto.Field( + proto.STRING, + number=10, + ) + end_time = proto.Field( + proto.STRING, + number=11, + ) + progress = proto.Field( + proto.MESSAGE, + number=12, + message="OperationProgress", + ) cluster_conditions = proto.RepeatedField( - proto.MESSAGE, number=13, message="StatusCondition", + proto.MESSAGE, + number=13, + message="StatusCondition", ) nodepool_conditions = proto.RepeatedField( - proto.MESSAGE, number=14, message="StatusCondition", + proto.MESSAGE, + number=14, + message="StatusCondition", + ) + error = proto.Field( + proto.MESSAGE, + number=15, + message=status_pb2.Status, ) - error = proto.Field(proto.MESSAGE, number=15, message=status_pb2.Status,) class OperationProgress(proto.Message): @@ -2013,15 +2615,45 @@ class Metric(proto.Message): This field is a member of `oneof`_ ``value``. """ - name = proto.Field(proto.STRING, number=1,) - int_value = proto.Field(proto.INT64, number=2, oneof="value",) - double_value = proto.Field(proto.DOUBLE, number=3, oneof="value",) - string_value = proto.Field(proto.STRING, number=4, oneof="value",) + name = proto.Field( + proto.STRING, + number=1, + ) + int_value = proto.Field( + proto.INT64, + number=2, + oneof="value", + ) + double_value = proto.Field( + proto.DOUBLE, + number=3, + oneof="value", + ) + string_value = proto.Field( + proto.STRING, + number=4, + oneof="value", + ) - name = proto.Field(proto.STRING, number=1,) - status = proto.Field(proto.ENUM, number=2, enum="Operation.Status",) - metrics = proto.RepeatedField(proto.MESSAGE, number=3, message=Metric,) - stages = proto.RepeatedField(proto.MESSAGE, number=4, message="OperationProgress",) + name = proto.Field( + proto.STRING, + number=1, + ) + status = proto.Field( + proto.ENUM, + number=2, + enum="Operation.Status", + ) + metrics = proto.RepeatedField( + proto.MESSAGE, + number=3, + message=Metric, + ) + stages = proto.RepeatedField( + proto.MESSAGE, + number=4, + message="OperationProgress", + ) class CreateClusterRequest(proto.Message): @@ -2047,10 +2679,23 @@ class CreateClusterRequest(proto.Message): created. Specified in the format ``projects/*/locations/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster = proto.Field(proto.MESSAGE, number=3, message="Cluster",) - parent = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster = proto.Field( + proto.MESSAGE, + number=3, + message="Cluster", + ) + parent = proto.Field( + proto.STRING, + number=5, + ) class GetClusterRequest(proto.Message): @@ -2078,10 +2723,22 @@ class GetClusterRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=5, + ) class UpdateClusterRequest(proto.Message): @@ -2111,11 +2768,27 @@ class UpdateClusterRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - update = proto.Field(proto.MESSAGE, number=4, message="ClusterUpdate",) - name = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + update = proto.Field( + proto.MESSAGE, + number=4, + message="ClusterUpdate", + ) + name = proto.Field( + proto.STRING, + number=5, + ) class UpdateNodePoolRequest(proto.Message): @@ -2187,26 +2860,68 @@ class UpdateNodePoolRequest(proto.Message): Enable or disable gvnic on the node pool. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - node_version = proto.Field(proto.STRING, number=5,) - image_type = proto.Field(proto.STRING, number=6,) - name = proto.Field(proto.STRING, number=8,) - locations = proto.RepeatedField(proto.STRING, number=13,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + node_version = proto.Field( + proto.STRING, + number=5, + ) + image_type = proto.Field( + proto.STRING, + number=6, + ) + name = proto.Field( + proto.STRING, + number=8, + ) + locations = proto.RepeatedField( + proto.STRING, + number=13, + ) workload_metadata_config = proto.Field( - proto.MESSAGE, number=14, message="WorkloadMetadataConfig", + proto.MESSAGE, + number=14, + message="WorkloadMetadataConfig", ) upgrade_settings = proto.Field( - proto.MESSAGE, number=15, message="NodePool.UpgradeSettings", + proto.MESSAGE, + number=15, + message="NodePool.UpgradeSettings", ) linux_node_config = proto.Field( - proto.MESSAGE, number=19, message="LinuxNodeConfig", + proto.MESSAGE, + number=19, + message="LinuxNodeConfig", + ) + kubelet_config = proto.Field( + proto.MESSAGE, + number=20, + message="NodeKubeletConfig", + ) + gcfs_config = proto.Field( + proto.MESSAGE, + number=22, + message="GcfsConfig", + ) + gvnic = proto.Field( + proto.MESSAGE, + number=29, + message="VirtualNIC", ) - kubelet_config = proto.Field(proto.MESSAGE, number=20, message="NodeKubeletConfig",) - gcfs_config = proto.Field(proto.MESSAGE, number=22, message="GcfsConfig",) - gvnic = proto.Field(proto.MESSAGE, number=29, message="VirtualNIC",) class SetNodePoolAutoscalingRequest(proto.Message): @@ -2242,12 +2957,31 @@ class SetNodePoolAutoscalingRequest(proto.Message): ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - autoscaling = proto.Field(proto.MESSAGE, number=5, message="NodePoolAutoscaling",) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + autoscaling = proto.Field( + proto.MESSAGE, + number=5, + message="NodePoolAutoscaling", + ) + name = proto.Field( + proto.STRING, + number=6, + ) class SetLoggingServiceRequest(proto.Message): @@ -2290,11 +3024,26 @@ class SetLoggingServiceRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - logging_service = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + logging_service = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=5, + ) class SetMonitoringServiceRequest(proto.Message): @@ -2338,11 +3087,26 @@ class SetMonitoringServiceRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - monitoring_service = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + monitoring_service = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class SetAddonsConfigRequest(proto.Message): @@ -2374,11 +3138,27 @@ class SetAddonsConfigRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - addons_config = proto.Field(proto.MESSAGE, number=4, message="AddonsConfig",) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + addons_config = proto.Field( + proto.MESSAGE, + number=4, + message="AddonsConfig", + ) + name = proto.Field( + proto.STRING, + number=6, + ) class SetLocationsRequest(proto.Message): @@ -2415,11 +3195,26 @@ class SetLocationsRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - locations = proto.RepeatedField(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + locations = proto.RepeatedField( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class UpdateMasterRequest(proto.Message): @@ -2460,11 +3255,26 @@ class UpdateMasterRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - master_version = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + master_version = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=7, + ) class SetMasterAuthRequest(proto.Message): @@ -2504,12 +3314,32 @@ class Action(proto.Enum): GENERATE_PASSWORD = 2 SET_USERNAME = 3 - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - action = proto.Field(proto.ENUM, number=4, enum=Action,) - update = proto.Field(proto.MESSAGE, number=5, message="MasterAuth",) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + action = proto.Field( + proto.ENUM, + number=4, + enum=Action, + ) + update = proto.Field( + proto.MESSAGE, + number=5, + message="MasterAuth", + ) + name = proto.Field( + proto.STRING, + number=7, + ) class DeleteClusterRequest(proto.Message): @@ -2537,10 +3367,22 @@ class DeleteClusterRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=4,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=4, + ) class ListClustersRequest(proto.Message): @@ -2564,9 +3406,18 @@ class ListClustersRequest(proto.Message): Location "-" matches all zones and all regions. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - parent = proto.Field(proto.STRING, number=4,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + parent = proto.Field( + proto.STRING, + number=4, + ) class ListClustersResponse(proto.Message): @@ -2581,8 +3432,15 @@ class ListClustersResponse(proto.Message): clusters returned may be missing those zones. """ - clusters = proto.RepeatedField(proto.MESSAGE, number=1, message="Cluster",) - missing_zones = proto.RepeatedField(proto.STRING, number=2,) + clusters = proto.RepeatedField( + proto.MESSAGE, + number=1, + message="Cluster", + ) + missing_zones = proto.RepeatedField( + proto.STRING, + number=2, + ) class GetOperationRequest(proto.Message): @@ -2610,10 +3468,22 @@ class GetOperationRequest(proto.Message): ``projects/*/locations/*/operations/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - operation_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + operation_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=5, + ) class ListOperationsRequest(proto.Message): @@ -2638,9 +3508,18 @@ class ListOperationsRequest(proto.Message): and all regions. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - parent = proto.Field(proto.STRING, number=4,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + parent = proto.Field( + proto.STRING, + number=4, + ) class CancelOperationRequest(proto.Message): @@ -2668,10 +3547,22 @@ class CancelOperationRequest(proto.Message): ``projects/*/locations/*/operations/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - operation_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=4,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + operation_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=4, + ) class ListOperationsResponse(proto.Message): @@ -2688,8 +3579,15 @@ class ListOperationsResponse(proto.Message): operations from those zones. """ - operations = proto.RepeatedField(proto.MESSAGE, number=1, message="Operation",) - missing_zones = proto.RepeatedField(proto.STRING, number=2,) + operations = proto.RepeatedField( + proto.MESSAGE, + number=1, + message="Operation", + ) + missing_zones = proto.RepeatedField( + proto.STRING, + number=2, + ) class GetServerConfigRequest(proto.Message): @@ -2712,9 +3610,18 @@ class GetServerConfigRequest(proto.Message): specified in the format ``projects/*/locations/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - name = proto.Field(proto.STRING, number=4,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + name = proto.Field( + proto.STRING, + number=4, + ) class ServerConfig(proto.Message): @@ -2753,17 +3660,44 @@ class ReleaseChannelConfig(proto.Message): List of valid versions for the channel. """ - channel = proto.Field(proto.ENUM, number=1, enum="ReleaseChannel.Channel",) - default_version = proto.Field(proto.STRING, number=2,) - valid_versions = proto.RepeatedField(proto.STRING, number=4,) + channel = proto.Field( + proto.ENUM, + number=1, + enum="ReleaseChannel.Channel", + ) + default_version = proto.Field( + proto.STRING, + number=2, + ) + valid_versions = proto.RepeatedField( + proto.STRING, + number=4, + ) - default_cluster_version = proto.Field(proto.STRING, number=1,) - valid_node_versions = proto.RepeatedField(proto.STRING, number=3,) - default_image_type = proto.Field(proto.STRING, number=4,) - valid_image_types = proto.RepeatedField(proto.STRING, number=5,) - valid_master_versions = proto.RepeatedField(proto.STRING, number=6,) + default_cluster_version = proto.Field( + proto.STRING, + number=1, + ) + valid_node_versions = proto.RepeatedField( + proto.STRING, + number=3, + ) + default_image_type = proto.Field( + proto.STRING, + number=4, + ) + valid_image_types = proto.RepeatedField( + proto.STRING, + number=5, + ) + valid_master_versions = proto.RepeatedField( + proto.STRING, + number=6, + ) channels = proto.RepeatedField( - proto.MESSAGE, number=9, message=ReleaseChannelConfig, + proto.MESSAGE, + number=9, + message=ReleaseChannelConfig, ) @@ -2794,11 +3728,27 @@ class CreateNodePoolRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool = proto.Field(proto.MESSAGE, number=4, message="NodePool",) - parent = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool = proto.Field( + proto.MESSAGE, + number=4, + message="NodePool", + ) + parent = proto.Field( + proto.STRING, + number=6, + ) class DeleteNodePoolRequest(proto.Message): @@ -2830,11 +3780,26 @@ class DeleteNodePoolRequest(proto.Message): ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class ListNodePoolsRequest(proto.Message): @@ -2862,10 +3827,22 @@ class ListNodePoolsRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - parent = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + parent = proto.Field( + proto.STRING, + number=5, + ) class GetNodePoolRequest(proto.Message): @@ -2897,11 +3874,26 @@ class GetNodePoolRequest(proto.Message): ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class NodePool(proto.Message): @@ -3023,29 +4015,87 @@ class UpgradeSettings(proto.Message): status is Ready. """ - max_surge = proto.Field(proto.INT32, number=1,) - max_unavailable = proto.Field(proto.INT32, number=2,) - - name = proto.Field(proto.STRING, number=1,) - config = proto.Field(proto.MESSAGE, number=2, message="NodeConfig",) - initial_node_count = proto.Field(proto.INT32, number=3,) - locations = proto.RepeatedField(proto.STRING, number=13,) - network_config = proto.Field(proto.MESSAGE, number=14, message="NodeNetworkConfig",) - self_link = proto.Field(proto.STRING, number=100,) - version = proto.Field(proto.STRING, number=101,) - instance_group_urls = proto.RepeatedField(proto.STRING, number=102,) - status = proto.Field(proto.ENUM, number=103, enum=Status,) - status_message = proto.Field(proto.STRING, number=104,) - autoscaling = proto.Field(proto.MESSAGE, number=4, message="NodePoolAutoscaling",) - management = proto.Field(proto.MESSAGE, number=5, message="NodeManagement",) + max_surge = proto.Field( + proto.INT32, + number=1, + ) + max_unavailable = proto.Field( + proto.INT32, + number=2, + ) + + name = proto.Field( + proto.STRING, + number=1, + ) + config = proto.Field( + proto.MESSAGE, + number=2, + message="NodeConfig", + ) + initial_node_count = proto.Field( + proto.INT32, + number=3, + ) + locations = proto.RepeatedField( + proto.STRING, + number=13, + ) + network_config = proto.Field( + proto.MESSAGE, + number=14, + message="NodeNetworkConfig", + ) + self_link = proto.Field( + proto.STRING, + number=100, + ) + version = proto.Field( + proto.STRING, + number=101, + ) + instance_group_urls = proto.RepeatedField( + proto.STRING, + number=102, + ) + status = proto.Field( + proto.ENUM, + number=103, + enum=Status, + ) + status_message = proto.Field( + proto.STRING, + number=104, + ) + autoscaling = proto.Field( + proto.MESSAGE, + number=4, + message="NodePoolAutoscaling", + ) + management = proto.Field( + proto.MESSAGE, + number=5, + message="NodeManagement", + ) max_pods_constraint = proto.Field( - proto.MESSAGE, number=6, message="MaxPodsConstraint", + proto.MESSAGE, + number=6, + message="MaxPodsConstraint", ) conditions = proto.RepeatedField( - proto.MESSAGE, number=105, message="StatusCondition", + proto.MESSAGE, + number=105, + message="StatusCondition", + ) + pod_ipv4_cidr_size = proto.Field( + proto.INT32, + number=7, + ) + upgrade_settings = proto.Field( + proto.MESSAGE, + number=107, + message=UpgradeSettings, ) - pod_ipv4_cidr_size = proto.Field(proto.INT32, number=7,) - upgrade_settings = proto.Field(proto.MESSAGE, number=107, message=UpgradeSettings,) class NodeManagement(proto.Message): @@ -3071,10 +4121,18 @@ class NodeManagement(proto.Message): pool. """ - auto_upgrade = proto.Field(proto.BOOL, number=1,) - auto_repair = proto.Field(proto.BOOL, number=2,) + auto_upgrade = proto.Field( + proto.BOOL, + number=1, + ) + auto_repair = proto.Field( + proto.BOOL, + number=2, + ) upgrade_options = proto.Field( - proto.MESSAGE, number=10, message="AutoUpgradeOptions", + proto.MESSAGE, + number=10, + message="AutoUpgradeOptions", ) @@ -3093,8 +4151,14 @@ class AutoUpgradeOptions(proto.Message): commence with the description of the upgrade. """ - auto_upgrade_start_time = proto.Field(proto.STRING, number=1,) - description = proto.Field(proto.STRING, number=2,) + auto_upgrade_start_time = proto.Field( + proto.STRING, + number=1, + ) + description = proto.Field( + proto.STRING, + number=2, + ) class MaintenancePolicy(proto.Message): @@ -3115,8 +4179,15 @@ class MaintenancePolicy(proto.Message): policy. """ - window = proto.Field(proto.MESSAGE, number=1, message="MaintenanceWindow",) - resource_version = proto.Field(proto.STRING, number=3,) + window = proto.Field( + proto.MESSAGE, + number=1, + message="MaintenanceWindow", + ) + resource_version = proto.Field( + proto.STRING, + number=3, + ) class MaintenanceWindow(proto.Message): @@ -3144,20 +4215,29 @@ class MaintenanceWindow(proto.Message): occur at any time. This field is a member of `oneof`_ ``policy``. - maintenance_exclusions (Sequence[google.cloud.container_v1.types.MaintenanceWindow.MaintenanceExclusionsEntry]): + maintenance_exclusions (Mapping[str, google.cloud.container_v1.types.TimeWindow]): Exceptions to maintenance window. Non-emergency maintenance should not occur in these windows. """ daily_maintenance_window = proto.Field( - proto.MESSAGE, number=2, oneof="policy", message="DailyMaintenanceWindow", + proto.MESSAGE, + number=2, + oneof="policy", + message="DailyMaintenanceWindow", ) recurring_window = proto.Field( - proto.MESSAGE, number=3, oneof="policy", message="RecurringTimeWindow", + proto.MESSAGE, + number=3, + oneof="policy", + message="RecurringTimeWindow", ) maintenance_exclusions = proto.MapField( - proto.STRING, proto.MESSAGE, number=4, message="TimeWindow", + proto.STRING, + proto.MESSAGE, + number=4, + message="TimeWindow", ) @@ -3180,10 +4260,21 @@ class TimeWindow(proto.Message): """ maintenance_exclusion_options = proto.Field( - proto.MESSAGE, number=3, oneof="options", message="MaintenanceExclusionOptions", + proto.MESSAGE, + number=3, + oneof="options", + message="MaintenanceExclusionOptions", + ) + start_time = proto.Field( + proto.MESSAGE, + number=1, + message=timestamp_pb2.Timestamp, + ) + end_time = proto.Field( + proto.MESSAGE, + number=2, + message=timestamp_pb2.Timestamp, ) - start_time = proto.Field(proto.MESSAGE, number=1, message=timestamp_pb2.Timestamp,) - end_time = proto.Field(proto.MESSAGE, number=2, message=timestamp_pb2.Timestamp,) class MaintenanceExclusionOptions(proto.Message): @@ -3201,7 +4292,11 @@ class Scope(proto.Enum): NO_MINOR_UPGRADES = 1 NO_MINOR_OR_NODE_UPGRADES = 2 - scope = proto.Field(proto.ENUM, number=1, enum=Scope,) + scope = proto.Field( + proto.ENUM, + number=1, + enum=Scope, + ) class RecurringTimeWindow(proto.Message): @@ -3251,8 +4346,15 @@ class RecurringTimeWindow(proto.Message): SECONDLY are not supported. """ - window = proto.Field(proto.MESSAGE, number=1, message="TimeWindow",) - recurrence = proto.Field(proto.STRING, number=2,) + window = proto.Field( + proto.MESSAGE, + number=1, + message="TimeWindow", + ) + recurrence = proto.Field( + proto.STRING, + number=2, + ) class DailyMaintenanceWindow(proto.Message): @@ -3272,8 +4374,14 @@ class DailyMaintenanceWindow(proto.Message): "PTnHnMnS". """ - start_time = proto.Field(proto.STRING, number=2,) - duration = proto.Field(proto.STRING, number=3,) + start_time = proto.Field( + proto.STRING, + number=2, + ) + duration = proto.Field( + proto.STRING, + number=3, + ) class SetNodePoolManagementRequest(proto.Message): @@ -3309,12 +4417,31 @@ class SetNodePoolManagementRequest(proto.Message): format ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - management = proto.Field(proto.MESSAGE, number=5, message="NodeManagement",) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + management = proto.Field( + proto.MESSAGE, + number=5, + message="NodeManagement", + ) + name = proto.Field( + proto.STRING, + number=7, + ) class SetNodePoolSizeRequest(proto.Message): @@ -3349,12 +4476,30 @@ class SetNodePoolSizeRequest(proto.Message): ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - node_count = proto.Field(proto.INT32, number=5,) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + node_count = proto.Field( + proto.INT32, + number=5, + ) + name = proto.Field( + proto.STRING, + number=7, + ) class RollbackNodePoolUpgradeRequest(proto.Message): @@ -3388,11 +4533,26 @@ class RollbackNodePoolUpgradeRequest(proto.Message): ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class ListNodePoolsResponse(proto.Message): @@ -3403,7 +4563,11 @@ class ListNodePoolsResponse(proto.Message): A list of node pools for a cluster. """ - node_pools = proto.RepeatedField(proto.MESSAGE, number=1, message="NodePool",) + node_pools = proto.RepeatedField( + proto.MESSAGE, + number=1, + message="NodePool", + ) class ClusterAutoscaling(proto.Message): @@ -3436,15 +4600,29 @@ class AutoscalingProfile(proto.Enum): OPTIMIZE_UTILIZATION = 1 BALANCED = 2 - enable_node_autoprovisioning = proto.Field(proto.BOOL, number=1,) + enable_node_autoprovisioning = proto.Field( + proto.BOOL, + number=1, + ) resource_limits = proto.RepeatedField( - proto.MESSAGE, number=2, message="ResourceLimit", + proto.MESSAGE, + number=2, + message="ResourceLimit", + ) + autoscaling_profile = proto.Field( + proto.ENUM, + number=3, + enum=AutoscalingProfile, ) - autoscaling_profile = proto.Field(proto.ENUM, number=3, enum=AutoscalingProfile,) autoprovisioning_node_pool_defaults = proto.Field( - proto.MESSAGE, number=4, message="AutoprovisioningNodePoolDefaults", + proto.MESSAGE, + number=4, + message="AutoprovisioningNodePoolDefaults", + ) + autoprovisioning_locations = proto.RepeatedField( + proto.STRING, + number=5, ) - autoprovisioning_locations = proto.RepeatedField(proto.STRING, number=5,) class AutoprovisioningNodePoolDefaults(proto.Message): @@ -3499,20 +4677,49 @@ class AutoprovisioningNodePoolDefaults(proto.Message): The image type to use for NAP created node. """ - oauth_scopes = proto.RepeatedField(proto.STRING, number=1,) - service_account = proto.Field(proto.STRING, number=2,) + oauth_scopes = proto.RepeatedField( + proto.STRING, + number=1, + ) + service_account = proto.Field( + proto.STRING, + number=2, + ) upgrade_settings = proto.Field( - proto.MESSAGE, number=3, message="NodePool.UpgradeSettings", + proto.MESSAGE, + number=3, + message="NodePool.UpgradeSettings", + ) + management = proto.Field( + proto.MESSAGE, + number=4, + message="NodeManagement", + ) + min_cpu_platform = proto.Field( + proto.STRING, + number=5, + ) + disk_size_gb = proto.Field( + proto.INT32, + number=6, + ) + disk_type = proto.Field( + proto.STRING, + number=7, ) - management = proto.Field(proto.MESSAGE, number=4, message="NodeManagement",) - min_cpu_platform = proto.Field(proto.STRING, number=5,) - disk_size_gb = proto.Field(proto.INT32, number=6,) - disk_type = proto.Field(proto.STRING, number=7,) shielded_instance_config = proto.Field( - proto.MESSAGE, number=8, message="ShieldedInstanceConfig", + proto.MESSAGE, + number=8, + message="ShieldedInstanceConfig", + ) + boot_disk_kms_key = proto.Field( + proto.STRING, + number=9, + ) + image_type = proto.Field( + proto.STRING, + number=10, ) - boot_disk_kms_key = proto.Field(proto.STRING, number=9,) - image_type = proto.Field(proto.STRING, number=10,) class ResourceLimit(proto.Message): @@ -3531,9 +4738,18 @@ class ResourceLimit(proto.Message): cluster. """ - resource_type = proto.Field(proto.STRING, number=1,) - minimum = proto.Field(proto.INT64, number=2,) - maximum = proto.Field(proto.INT64, number=3,) + resource_type = proto.Field( + proto.STRING, + number=1, + ) + minimum = proto.Field( + proto.INT64, + number=2, + ) + maximum = proto.Field( + proto.INT64, + number=3, + ) class NodePoolAutoscaling(proto.Message): @@ -3555,10 +4771,22 @@ class NodePoolAutoscaling(proto.Message): Can this node pool be deleted automatically. """ - enabled = proto.Field(proto.BOOL, number=1,) - min_node_count = proto.Field(proto.INT32, number=2,) - max_node_count = proto.Field(proto.INT32, number=3,) - autoprovisioned = proto.Field(proto.BOOL, number=4,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) + min_node_count = proto.Field( + proto.INT32, + number=2, + ) + max_node_count = proto.Field( + proto.INT32, + number=3, + ) + autoprovisioned = proto.Field( + proto.BOOL, + number=4, + ) class SetLabelsRequest(proto.Message): @@ -3582,7 +4810,7 @@ class SetLabelsRequest(proto.Message): Deprecated. The name of the cluster. This field has been deprecated and replaced by the name field. - resource_labels (Sequence[google.cloud.container_v1.types.SetLabelsRequest.ResourceLabelsEntry]): + resource_labels (Mapping[str, str]): Required. The labels to set for that cluster. label_fingerprint (str): Required. The fingerprint of the previous set of labels for @@ -3598,12 +4826,31 @@ class SetLabelsRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - resource_labels = proto.MapField(proto.STRING, proto.STRING, number=4,) - label_fingerprint = proto.Field(proto.STRING, number=5,) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + resource_labels = proto.MapField( + proto.STRING, + proto.STRING, + number=4, + ) + label_fingerprint = proto.Field( + proto.STRING, + number=5, + ) + name = proto.Field( + proto.STRING, + number=7, + ) class SetLegacyAbacRequest(proto.Message): @@ -3635,11 +4882,26 @@ class SetLegacyAbacRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - enabled = proto.Field(proto.BOOL, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + enabled = proto.Field( + proto.BOOL, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class StartIPRotationRequest(proto.Message): @@ -3672,11 +4934,26 @@ class StartIPRotationRequest(proto.Message): rotation. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=6,) - rotate_credentials = proto.Field(proto.BOOL, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=6, + ) + rotate_credentials = proto.Field( + proto.BOOL, + number=7, + ) class CompleteIPRotationRequest(proto.Message): @@ -3705,10 +4982,22 @@ class CompleteIPRotationRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=7, + ) class AcceleratorConfig(proto.Message): @@ -3728,9 +5017,18 @@ class AcceleratorConfig(proto.Message): guide `__. """ - accelerator_count = proto.Field(proto.INT64, number=1,) - accelerator_type = proto.Field(proto.STRING, number=2,) - gpu_partition_size = proto.Field(proto.STRING, number=3,) + accelerator_count = proto.Field( + proto.INT64, + number=1, + ) + accelerator_type = proto.Field( + proto.STRING, + number=2, + ) + gpu_partition_size = proto.Field( + proto.STRING, + number=3, + ) class WorkloadMetadataConfig(proto.Message): @@ -3751,7 +5049,11 @@ class Mode(proto.Enum): GCE_METADATA = 1 GKE_METADATA = 2 - mode = proto.Field(proto.ENUM, number=2, enum=Mode,) + mode = proto.Field( + proto.ENUM, + number=2, + enum=Mode, + ) class SetNetworkPolicyRequest(proto.Message): @@ -3783,11 +5085,27 @@ class SetNetworkPolicyRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - network_policy = proto.Field(proto.MESSAGE, number=4, message="NetworkPolicy",) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + network_policy = proto.Field( + proto.MESSAGE, + number=4, + message="NetworkPolicy", + ) + name = proto.Field( + proto.STRING, + number=6, + ) class SetMaintenancePolicyRequest(proto.Message): @@ -3815,13 +5133,27 @@ class SetMaintenancePolicyRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) maintenance_policy = proto.Field( - proto.MESSAGE, number=4, message="MaintenancePolicy", + proto.MESSAGE, + number=4, + message="MaintenancePolicy", + ) + name = proto.Field( + proto.STRING, + number=5, ) - name = proto.Field(proto.STRING, number=5,) class StatusCondition(proto.Message): @@ -3849,9 +5181,20 @@ class Code(proto.Enum): CLOUD_KMS_KEY_ERROR = 7 CA_EXPIRING = 9 - code = proto.Field(proto.ENUM, number=1, enum=Code,) - message = proto.Field(proto.STRING, number=2,) - canonical_code = proto.Field(proto.ENUM, number=3, enum=code_pb2.Code,) + code = proto.Field( + proto.ENUM, + number=1, + enum=Code, + ) + message = proto.Field( + proto.STRING, + number=2, + ) + canonical_code = proto.Field( + proto.ENUM, + number=3, + enum=code_pb2.Code, + ) class NetworkConfig(proto.Message): @@ -3900,20 +5243,46 @@ class NetworkConfig(proto.Message): not. """ - network = proto.Field(proto.STRING, number=1,) - subnetwork = proto.Field(proto.STRING, number=2,) - enable_intra_node_visibility = proto.Field(proto.BOOL, number=5,) + network = proto.Field( + proto.STRING, + number=1, + ) + subnetwork = proto.Field( + proto.STRING, + number=2, + ) + enable_intra_node_visibility = proto.Field( + proto.BOOL, + number=5, + ) default_snat_status = proto.Field( - proto.MESSAGE, number=7, message="DefaultSnatStatus", + proto.MESSAGE, + number=7, + message="DefaultSnatStatus", + ) + enable_l4ilb_subsetting = proto.Field( + proto.BOOL, + number=10, + ) + datapath_provider = proto.Field( + proto.ENUM, + number=11, + enum="DatapathProvider", ) - enable_l4ilb_subsetting = proto.Field(proto.BOOL, number=10,) - datapath_provider = proto.Field(proto.ENUM, number=11, enum="DatapathProvider",) private_ipv6_google_access = proto.Field( - proto.ENUM, number=12, enum="PrivateIPv6GoogleAccess", + proto.ENUM, + number=12, + enum="PrivateIPv6GoogleAccess", + ) + dns_config = proto.Field( + proto.MESSAGE, + number=13, + message="DNSConfig", ) - dns_config = proto.Field(proto.MESSAGE, number=13, message="DNSConfig",) service_external_ips_config = proto.Field( - proto.MESSAGE, number=15, message="ServiceExternalIPsConfig", + proto.MESSAGE, + number=15, + message="ServiceExternalIPsConfig", ) @@ -3926,7 +5295,10 @@ class ServiceExternalIPsConfig(proto.Message): allowed or not. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class GetOpenIDConfigRequest(proto.Message): @@ -3941,7 +5313,10 @@ class GetOpenIDConfigRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - parent = proto.Field(proto.STRING, number=1,) + parent = proto.Field( + proto.STRING, + number=1, + ) class GetOpenIDConfigResponse(proto.Message): @@ -3966,13 +5341,34 @@ class GetOpenIDConfigResponse(proto.Message): Supported grant types. """ - issuer = proto.Field(proto.STRING, number=1,) - jwks_uri = proto.Field(proto.STRING, number=2,) - response_types_supported = proto.RepeatedField(proto.STRING, number=3,) - subject_types_supported = proto.RepeatedField(proto.STRING, number=4,) - id_token_signing_alg_values_supported = proto.RepeatedField(proto.STRING, number=5,) - claims_supported = proto.RepeatedField(proto.STRING, number=6,) - grant_types = proto.RepeatedField(proto.STRING, number=7,) + issuer = proto.Field( + proto.STRING, + number=1, + ) + jwks_uri = proto.Field( + proto.STRING, + number=2, + ) + response_types_supported = proto.RepeatedField( + proto.STRING, + number=3, + ) + subject_types_supported = proto.RepeatedField( + proto.STRING, + number=4, + ) + id_token_signing_alg_values_supported = proto.RepeatedField( + proto.STRING, + number=5, + ) + claims_supported = proto.RepeatedField( + proto.STRING, + number=6, + ) + grant_types = proto.RepeatedField( + proto.STRING, + number=7, + ) class GetJSONWebKeysRequest(proto.Message): @@ -3988,7 +5384,10 @@ class GetJSONWebKeysRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - parent = proto.Field(proto.STRING, number=1,) + parent = proto.Field( + proto.STRING, + number=1, + ) class Jwk(proto.Message): @@ -4015,15 +5414,42 @@ class Jwk(proto.Message): Used for ECDSA keys. """ - kty = proto.Field(proto.STRING, number=1,) - alg = proto.Field(proto.STRING, number=2,) - use = proto.Field(proto.STRING, number=3,) - kid = proto.Field(proto.STRING, number=4,) - n = proto.Field(proto.STRING, number=5,) - e = proto.Field(proto.STRING, number=6,) - x = proto.Field(proto.STRING, number=7,) - y = proto.Field(proto.STRING, number=8,) - crv = proto.Field(proto.STRING, number=9,) + kty = proto.Field( + proto.STRING, + number=1, + ) + alg = proto.Field( + proto.STRING, + number=2, + ) + use = proto.Field( + proto.STRING, + number=3, + ) + kid = proto.Field( + proto.STRING, + number=4, + ) + n = proto.Field( + proto.STRING, + number=5, + ) + e = proto.Field( + proto.STRING, + number=6, + ) + x = proto.Field( + proto.STRING, + number=7, + ) + y = proto.Field( + proto.STRING, + number=8, + ) + crv = proto.Field( + proto.STRING, + number=9, + ) class GetJSONWebKeysResponse(proto.Message): @@ -4036,7 +5462,11 @@ class GetJSONWebKeysResponse(proto.Message): cluster to sign token requests. """ - keys = proto.RepeatedField(proto.MESSAGE, number=1, message="Jwk",) + keys = proto.RepeatedField( + proto.MESSAGE, + number=1, + message="Jwk", + ) class ReleaseChannel(proto.Message): @@ -4059,7 +5489,11 @@ class Channel(proto.Enum): REGULAR = 2 STABLE = 3 - channel = proto.Field(proto.ENUM, number=1, enum=Channel,) + channel = proto.Field( + proto.ENUM, + number=1, + enum=Channel, + ) class IntraNodeVisibilityConfig(proto.Message): @@ -4072,7 +5506,10 @@ class IntraNodeVisibilityConfig(proto.Message): cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class ILBSubsettingConfig(proto.Message): @@ -4084,7 +5521,10 @@ class ILBSubsettingConfig(proto.Message): Enables l4 ILB subsetting for this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class DNSConfig(proto.Message): @@ -4116,9 +5556,20 @@ class DNSScope(proto.Enum): DNS_SCOPE_UNSPECIFIED = 0 VPC_SCOPE = 2 - cluster_dns = proto.Field(proto.ENUM, number=1, enum=Provider,) - cluster_dns_scope = proto.Field(proto.ENUM, number=2, enum=DNSScope,) - cluster_dns_domain = proto.Field(proto.STRING, number=3,) + cluster_dns = proto.Field( + proto.ENUM, + number=1, + enum=Provider, + ) + cluster_dns_scope = proto.Field( + proto.ENUM, + number=2, + enum=DNSScope, + ) + cluster_dns_domain = proto.Field( + proto.STRING, + number=3, + ) class MaxPodsConstraint(proto.Message): @@ -4130,7 +5581,10 @@ class MaxPodsConstraint(proto.Message): per node. """ - max_pods_per_node = proto.Field(proto.INT64, number=1,) + max_pods_per_node = proto.Field( + proto.INT64, + number=1, + ) class WorkloadIdentityConfig(proto.Message): @@ -4143,7 +5597,10 @@ class WorkloadIdentityConfig(proto.Message): service accounts to. """ - workload_pool = proto.Field(proto.STRING, number=2,) + workload_pool = proto.Field( + proto.STRING, + number=2, + ) class MeshCertificates(proto.Message): @@ -4166,7 +5623,9 @@ class MeshCertificates(proto.Message): """ enable_certificates = proto.Field( - proto.MESSAGE, number=1, message=wrappers_pb2.BoolValue, + proto.MESSAGE, + number=1, + message=wrappers_pb2.BoolValue, ) @@ -4188,8 +5647,15 @@ class State(proto.Enum): ENCRYPTED = 1 DECRYPTED = 2 - state = proto.Field(proto.ENUM, number=2, enum=State,) - key_name = proto.Field(proto.STRING, number=1,) + state = proto.Field( + proto.ENUM, + number=2, + enum=State, + ) + key_name = proto.Field( + proto.STRING, + number=1, + ) class ListUsableSubnetworksRequest(proto.Message): @@ -4219,10 +5685,22 @@ class ListUsableSubnetworksRequest(proto.Message): requests to get the next page of results. """ - parent = proto.Field(proto.STRING, number=1,) - filter = proto.Field(proto.STRING, number=2,) - page_size = proto.Field(proto.INT32, number=3,) - page_token = proto.Field(proto.STRING, number=4,) + parent = proto.Field( + proto.STRING, + number=1, + ) + filter = proto.Field( + proto.STRING, + number=2, + ) + page_size = proto.Field( + proto.INT32, + number=3, + ) + page_token = proto.Field( + proto.STRING, + number=4, + ) class ListUsableSubnetworksResponse(proto.Message): @@ -4246,9 +5724,14 @@ def raw_page(self): return self subnetworks = proto.RepeatedField( - proto.MESSAGE, number=1, message="UsableSubnetwork", + proto.MESSAGE, + number=1, + message="UsableSubnetwork", + ) + next_page_token = proto.Field( + proto.STRING, + number=2, ) - next_page_token = proto.Field(proto.STRING, number=2,) class UsableSubnetworkSecondaryRange(proto.Message): @@ -4275,9 +5758,19 @@ class Status(proto.Enum): IN_USE_SHAREABLE_POD = 3 IN_USE_MANAGED_POD = 4 - range_name = proto.Field(proto.STRING, number=1,) - ip_cidr_range = proto.Field(proto.STRING, number=2,) - status = proto.Field(proto.ENUM, number=3, enum=Status,) + range_name = proto.Field( + proto.STRING, + number=1, + ) + ip_cidr_range = proto.Field( + proto.STRING, + number=2, + ) + status = proto.Field( + proto.ENUM, + number=3, + enum=Status, + ) class UsableSubnetwork(proto.Message): @@ -4306,13 +5799,27 @@ class UsableSubnetwork(proto.Message): message will be given by status_message. """ - subnetwork = proto.Field(proto.STRING, number=1,) - network = proto.Field(proto.STRING, number=2,) - ip_cidr_range = proto.Field(proto.STRING, number=3,) + subnetwork = proto.Field( + proto.STRING, + number=1, + ) + network = proto.Field( + proto.STRING, + number=2, + ) + ip_cidr_range = proto.Field( + proto.STRING, + number=3, + ) secondary_ip_ranges = proto.RepeatedField( - proto.MESSAGE, number=4, message="UsableSubnetworkSecondaryRange", + proto.MESSAGE, + number=4, + message="UsableSubnetworkSecondaryRange", + ) + status_message = proto.Field( + proto.STRING, + number=5, ) - status_message = proto.Field(proto.STRING, number=5,) class ResourceUsageExportConfig(proto.Message): @@ -4341,7 +5848,10 @@ class BigQueryDestination(proto.Message): The ID of a BigQuery Dataset. """ - dataset_id = proto.Field(proto.STRING, number=1,) + dataset_id = proto.Field( + proto.STRING, + number=1, + ) class ConsumptionMeteringConfig(proto.Message): r"""Parameters for controlling consumption metering. @@ -4354,14 +5864,24 @@ class ConsumptionMeteringConfig(proto.Message): consumption records. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) bigquery_destination = proto.Field( - proto.MESSAGE, number=1, message=BigQueryDestination, + proto.MESSAGE, + number=1, + message=BigQueryDestination, + ) + enable_network_egress_metering = proto.Field( + proto.BOOL, + number=2, ) - enable_network_egress_metering = proto.Field(proto.BOOL, number=2,) consumption_metering_config = proto.Field( - proto.MESSAGE, number=3, message=ConsumptionMeteringConfig, + proto.MESSAGE, + number=3, + message=ConsumptionMeteringConfig, ) @@ -4375,7 +5895,10 @@ class VerticalPodAutoscaling(proto.Message): Enables vertical pod autoscaling. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class DefaultSnatStatus(proto.Message): @@ -4387,7 +5910,10 @@ class DefaultSnatStatus(proto.Message): Disables cluster default sNAT rules. """ - disabled = proto.Field(proto.BOOL, number=1,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) class ShieldedNodes(proto.Message): @@ -4399,7 +5925,10 @@ class ShieldedNodes(proto.Message): on all nodes in this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class VirtualNIC(proto.Message): @@ -4411,7 +5940,10 @@ class VirtualNIC(proto.Message): node pool. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class NotificationConfig(proto.Message): @@ -4448,10 +5980,18 @@ class PubSub(proto.Message): event types will be sent """ - enabled = proto.Field(proto.BOOL, number=1,) - topic = proto.Field(proto.STRING, number=2,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) + topic = proto.Field( + proto.STRING, + number=2, + ) filter = proto.Field( - proto.MESSAGE, number=3, message="NotificationConfig.Filter", + proto.MESSAGE, + number=3, + message="NotificationConfig.Filter", ) class Filter(proto.Message): @@ -4467,10 +6007,16 @@ class Filter(proto.Message): """ event_type = proto.RepeatedField( - proto.ENUM, number=1, enum="NotificationConfig.EventType", + proto.ENUM, + number=1, + enum="NotificationConfig.EventType", ) - pubsub = proto.Field(proto.MESSAGE, number=1, message=PubSub,) + pubsub = proto.Field( + proto.MESSAGE, + number=1, + message=PubSub, + ) class ConfidentialNodes(proto.Message): @@ -4483,7 +6029,10 @@ class ConfidentialNodes(proto.Message): for all nodes in this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class UpgradeEvent(proto.Message): @@ -4507,14 +6056,32 @@ class UpgradeEvent(proto.Message): of the node pool. """ - resource_type = proto.Field(proto.ENUM, number=1, enum="UpgradeResourceType",) - operation = proto.Field(proto.STRING, number=2,) + resource_type = proto.Field( + proto.ENUM, + number=1, + enum="UpgradeResourceType", + ) + operation = proto.Field( + proto.STRING, + number=2, + ) operation_start_time = proto.Field( - proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, + proto.MESSAGE, + number=3, + message=timestamp_pb2.Timestamp, + ) + current_version = proto.Field( + proto.STRING, + number=4, + ) + target_version = proto.Field( + proto.STRING, + number=5, + ) + resource = proto.Field( + proto.STRING, + number=6, ) - current_version = proto.Field(proto.STRING, number=4,) - target_version = proto.Field(proto.STRING, number=5,) - resource = proto.Field(proto.STRING, number=6,) class UpgradeAvailableEvent(proto.Message): @@ -4534,10 +6101,24 @@ class UpgradeAvailableEvent(proto.Message): example, the relative path of the node pool. """ - version = proto.Field(proto.STRING, number=1,) - resource_type = proto.Field(proto.ENUM, number=2, enum="UpgradeResourceType",) - release_channel = proto.Field(proto.MESSAGE, number=3, message="ReleaseChannel",) - resource = proto.Field(proto.STRING, number=4,) + version = proto.Field( + proto.STRING, + number=1, + ) + resource_type = proto.Field( + proto.ENUM, + number=2, + enum="UpgradeResourceType", + ) + release_channel = proto.Field( + proto.MESSAGE, + number=3, + message="ReleaseChannel", + ) + resource = proto.Field( + proto.STRING, + number=4, + ) class SecurityBulletinEvent(proto.Message): @@ -4585,16 +6166,46 @@ class SecurityBulletinEvent(proto.Message): their clusters safe. """ - resource_type_affected = proto.Field(proto.STRING, number=1,) - bulletin_id = proto.Field(proto.STRING, number=2,) - cve_ids = proto.RepeatedField(proto.STRING, number=3,) - severity = proto.Field(proto.STRING, number=4,) - bulletin_uri = proto.Field(proto.STRING, number=5,) - brief_description = proto.Field(proto.STRING, number=6,) - affected_supported_minors = proto.RepeatedField(proto.STRING, number=7,) - patched_versions = proto.RepeatedField(proto.STRING, number=8,) - suggested_upgrade_target = proto.Field(proto.STRING, number=9,) - manual_steps_required = proto.Field(proto.BOOL, number=10,) + resource_type_affected = proto.Field( + proto.STRING, + number=1, + ) + bulletin_id = proto.Field( + proto.STRING, + number=2, + ) + cve_ids = proto.RepeatedField( + proto.STRING, + number=3, + ) + severity = proto.Field( + proto.STRING, + number=4, + ) + bulletin_uri = proto.Field( + proto.STRING, + number=5, + ) + brief_description = proto.Field( + proto.STRING, + number=6, + ) + affected_supported_minors = proto.RepeatedField( + proto.STRING, + number=7, + ) + patched_versions = proto.RepeatedField( + proto.STRING, + number=8, + ) + suggested_upgrade_target = proto.Field( + proto.STRING, + number=9, + ) + manual_steps_required = proto.Field( + proto.BOOL, + number=10, + ) class Autopilot(proto.Message): @@ -4606,7 +6217,10 @@ class Autopilot(proto.Message): Enable Autopilot """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class LoggingConfig(proto.Message): @@ -4618,7 +6232,9 @@ class LoggingConfig(proto.Message): """ component_config = proto.Field( - proto.MESSAGE, number=1, message="LoggingComponentConfig", + proto.MESSAGE, + number=1, + message="LoggingComponentConfig", ) @@ -4638,7 +6254,11 @@ class Component(proto.Enum): SYSTEM_COMPONENTS = 1 WORKLOADS = 2 - enable_components = proto.RepeatedField(proto.ENUM, number=1, enum=Component,) + enable_components = proto.RepeatedField( + proto.ENUM, + number=1, + enum=Component, + ) class MonitoringConfig(proto.Message): @@ -4650,7 +6270,9 @@ class MonitoringConfig(proto.Message): """ component_config = proto.Field( - proto.MESSAGE, number=1, message="MonitoringComponentConfig", + proto.MESSAGE, + number=1, + message="MonitoringComponentConfig", ) @@ -4669,7 +6291,11 @@ class Component(proto.Enum): COMPONENT_UNSPECIFIED = 0 SYSTEM_COMPONENTS = 1 - enable_components = proto.RepeatedField(proto.ENUM, number=1, enum=Component,) + enable_components = proto.RepeatedField( + proto.ENUM, + number=1, + enum=Component, + ) __all__ = tuple(sorted(__protobuf__.manifest)) diff --git a/google/cloud/container_v1beta1/__init__.py b/google/cloud/container_v1beta1/__init__.py index b8454da5..fd68ddf7 100644 --- a/google/cloud/container_v1beta1/__init__.py +++ b/google/cloud/container_v1beta1/__init__.py @@ -14,150 +14,150 @@ # limitations under the License. # -from .services.cluster_manager import ClusterManagerClient -from .services.cluster_manager import ClusterManagerAsyncClient - -from .types.cluster_service import AcceleratorConfig -from .types.cluster_service import AddonsConfig -from .types.cluster_service import AdvancedMachineFeatures -from .types.cluster_service import AuthenticatorGroupsConfig -from .types.cluster_service import Autopilot -from .types.cluster_service import AutoprovisioningNodePoolDefaults -from .types.cluster_service import AutoUpgradeOptions -from .types.cluster_service import BinaryAuthorization -from .types.cluster_service import CancelOperationRequest -from .types.cluster_service import ClientCertificateConfig -from .types.cluster_service import CloudRunConfig -from .types.cluster_service import Cluster -from .types.cluster_service import ClusterAutoscaling -from .types.cluster_service import ClusterTelemetry -from .types.cluster_service import ClusterUpdate -from .types.cluster_service import CompleteIPRotationRequest -from .types.cluster_service import ConfidentialNodes -from .types.cluster_service import ConfigConnectorConfig -from .types.cluster_service import CreateClusterRequest -from .types.cluster_service import CreateNodePoolRequest -from .types.cluster_service import DailyMaintenanceWindow -from .types.cluster_service import DatabaseEncryption -from .types.cluster_service import DefaultSnatStatus -from .types.cluster_service import DeleteClusterRequest -from .types.cluster_service import DeleteNodePoolRequest -from .types.cluster_service import DnsCacheConfig -from .types.cluster_service import DNSConfig -from .types.cluster_service import EphemeralStorageConfig -from .types.cluster_service import GcePersistentDiskCsiDriverConfig -from .types.cluster_service import GcfsConfig -from .types.cluster_service import GcpFilestoreCsiDriverConfig -from .types.cluster_service import GetClusterRequest -from .types.cluster_service import GetJSONWebKeysRequest -from .types.cluster_service import GetJSONWebKeysResponse -from .types.cluster_service import GetNodePoolRequest -from .types.cluster_service import GetOpenIDConfigRequest -from .types.cluster_service import GetOpenIDConfigResponse -from .types.cluster_service import GetOperationRequest -from .types.cluster_service import GetServerConfigRequest -from .types.cluster_service import GkeBackupAgentConfig -from .types.cluster_service import HorizontalPodAutoscaling -from .types.cluster_service import HttpLoadBalancing -from .types.cluster_service import IdentityServiceConfig -from .types.cluster_service import ILBSubsettingConfig -from .types.cluster_service import IntraNodeVisibilityConfig -from .types.cluster_service import IPAllocationPolicy -from .types.cluster_service import IstioConfig -from .types.cluster_service import Jwk -from .types.cluster_service import KalmConfig -from .types.cluster_service import KubernetesDashboard -from .types.cluster_service import LegacyAbac -from .types.cluster_service import LinuxNodeConfig -from .types.cluster_service import ListClustersRequest -from .types.cluster_service import ListClustersResponse -from .types.cluster_service import ListLocationsRequest -from .types.cluster_service import ListLocationsResponse -from .types.cluster_service import ListNodePoolsRequest -from .types.cluster_service import ListNodePoolsResponse -from .types.cluster_service import ListOperationsRequest -from .types.cluster_service import ListOperationsResponse -from .types.cluster_service import ListUsableSubnetworksRequest -from .types.cluster_service import ListUsableSubnetworksResponse -from .types.cluster_service import Location -from .types.cluster_service import LoggingComponentConfig -from .types.cluster_service import LoggingConfig -from .types.cluster_service import MaintenanceExclusionOptions -from .types.cluster_service import MaintenancePolicy -from .types.cluster_service import MaintenanceWindow -from .types.cluster_service import ManagedPrometheusConfig -from .types.cluster_service import Master -from .types.cluster_service import MasterAuth -from .types.cluster_service import MasterAuthorizedNetworksConfig -from .types.cluster_service import MaxPodsConstraint -from .types.cluster_service import MeshCertificates -from .types.cluster_service import MonitoringComponentConfig -from .types.cluster_service import MonitoringConfig -from .types.cluster_service import NetworkConfig -from .types.cluster_service import NetworkPolicy -from .types.cluster_service import NetworkPolicyConfig -from .types.cluster_service import NetworkTags -from .types.cluster_service import NodeConfig -from .types.cluster_service import NodeConfigDefaults -from .types.cluster_service import NodeKubeletConfig -from .types.cluster_service import NodeLabels -from .types.cluster_service import NodeManagement -from .types.cluster_service import NodeNetworkConfig -from .types.cluster_service import NodePool -from .types.cluster_service import NodePoolAutoscaling -from .types.cluster_service import NodePoolDefaults -from .types.cluster_service import NodeTaint -from .types.cluster_service import NodeTaints -from .types.cluster_service import NotificationConfig -from .types.cluster_service import Operation -from .types.cluster_service import OperationProgress -from .types.cluster_service import PodSecurityPolicyConfig -from .types.cluster_service import PrivateClusterConfig -from .types.cluster_service import PrivateClusterMasterGlobalAccessConfig -from .types.cluster_service import RecurringTimeWindow -from .types.cluster_service import ReleaseChannel -from .types.cluster_service import ReservationAffinity -from .types.cluster_service import ResourceLimit -from .types.cluster_service import ResourceUsageExportConfig -from .types.cluster_service import RollbackNodePoolUpgradeRequest -from .types.cluster_service import SandboxConfig -from .types.cluster_service import SecurityBulletinEvent -from .types.cluster_service import ServerConfig -from .types.cluster_service import ServiceExternalIPsConfig -from .types.cluster_service import SetAddonsConfigRequest -from .types.cluster_service import SetLabelsRequest -from .types.cluster_service import SetLegacyAbacRequest -from .types.cluster_service import SetLocationsRequest -from .types.cluster_service import SetLoggingServiceRequest -from .types.cluster_service import SetMaintenancePolicyRequest -from .types.cluster_service import SetMasterAuthRequest -from .types.cluster_service import SetMonitoringServiceRequest -from .types.cluster_service import SetNetworkPolicyRequest -from .types.cluster_service import SetNodePoolAutoscalingRequest -from .types.cluster_service import SetNodePoolManagementRequest -from .types.cluster_service import SetNodePoolSizeRequest -from .types.cluster_service import ShieldedInstanceConfig -from .types.cluster_service import ShieldedNodes -from .types.cluster_service import StartIPRotationRequest -from .types.cluster_service import StatusCondition -from .types.cluster_service import TimeWindow -from .types.cluster_service import TpuConfig -from .types.cluster_service import UpdateClusterRequest -from .types.cluster_service import UpdateMasterRequest -from .types.cluster_service import UpdateNodePoolRequest -from .types.cluster_service import UpgradeAvailableEvent -from .types.cluster_service import UpgradeEvent -from .types.cluster_service import UsableSubnetwork -from .types.cluster_service import UsableSubnetworkSecondaryRange -from .types.cluster_service import VerticalPodAutoscaling -from .types.cluster_service import VirtualNIC -from .types.cluster_service import WindowsVersions -from .types.cluster_service import WorkloadCertificates -from .types.cluster_service import WorkloadIdentityConfig -from .types.cluster_service import WorkloadMetadataConfig -from .types.cluster_service import DatapathProvider -from .types.cluster_service import PrivateIPv6GoogleAccess -from .types.cluster_service import UpgradeResourceType +from .services.cluster_manager import ClusterManagerAsyncClient, ClusterManagerClient +from .types.cluster_service import ( + AcceleratorConfig, + AddonsConfig, + AdvancedMachineFeatures, + AuthenticatorGroupsConfig, + Autopilot, + AutoprovisioningNodePoolDefaults, + AutoUpgradeOptions, + BinaryAuthorization, + CancelOperationRequest, + ClientCertificateConfig, + CloudRunConfig, + Cluster, + ClusterAutoscaling, + ClusterTelemetry, + ClusterUpdate, + CompleteIPRotationRequest, + ConfidentialNodes, + ConfigConnectorConfig, + CreateClusterRequest, + CreateNodePoolRequest, + DailyMaintenanceWindow, + DatabaseEncryption, + DatapathProvider, + DefaultSnatStatus, + DeleteClusterRequest, + DeleteNodePoolRequest, + DnsCacheConfig, + DNSConfig, + EphemeralStorageConfig, + GcePersistentDiskCsiDriverConfig, + GcfsConfig, + GcpFilestoreCsiDriverConfig, + GetClusterRequest, + GetJSONWebKeysRequest, + GetJSONWebKeysResponse, + GetNodePoolRequest, + GetOpenIDConfigRequest, + GetOpenIDConfigResponse, + GetOperationRequest, + GetServerConfigRequest, + GkeBackupAgentConfig, + HorizontalPodAutoscaling, + HttpLoadBalancing, + IdentityServiceConfig, + ILBSubsettingConfig, + IntraNodeVisibilityConfig, + IPAllocationPolicy, + IstioConfig, + Jwk, + KalmConfig, + KubernetesDashboard, + LegacyAbac, + LinuxNodeConfig, + ListClustersRequest, + ListClustersResponse, + ListLocationsRequest, + ListLocationsResponse, + ListNodePoolsRequest, + ListNodePoolsResponse, + ListOperationsRequest, + ListOperationsResponse, + ListUsableSubnetworksRequest, + ListUsableSubnetworksResponse, + Location, + LoggingComponentConfig, + LoggingConfig, + MaintenanceExclusionOptions, + MaintenancePolicy, + MaintenanceWindow, + ManagedPrometheusConfig, + Master, + MasterAuth, + MasterAuthorizedNetworksConfig, + MaxPodsConstraint, + MeshCertificates, + MonitoringComponentConfig, + MonitoringConfig, + NetworkConfig, + NetworkPolicy, + NetworkPolicyConfig, + NetworkTags, + NodeConfig, + NodeConfigDefaults, + NodeKubeletConfig, + NodeLabels, + NodeManagement, + NodeNetworkConfig, + NodePool, + NodePoolAutoscaling, + NodePoolDefaults, + NodeTaint, + NodeTaints, + NotificationConfig, + Operation, + OperationProgress, + PodSecurityPolicyConfig, + PrivateClusterConfig, + PrivateClusterMasterGlobalAccessConfig, + PrivateIPv6GoogleAccess, + RecurringTimeWindow, + ReleaseChannel, + ReservationAffinity, + ResourceLimit, + ResourceUsageExportConfig, + RollbackNodePoolUpgradeRequest, + SandboxConfig, + SecurityBulletinEvent, + ServerConfig, + ServiceExternalIPsConfig, + SetAddonsConfigRequest, + SetLabelsRequest, + SetLegacyAbacRequest, + SetLocationsRequest, + SetLoggingServiceRequest, + SetMaintenancePolicyRequest, + SetMasterAuthRequest, + SetMonitoringServiceRequest, + SetNetworkPolicyRequest, + SetNodePoolAutoscalingRequest, + SetNodePoolManagementRequest, + SetNodePoolSizeRequest, + ShieldedInstanceConfig, + ShieldedNodes, + StartIPRotationRequest, + StatusCondition, + TimeWindow, + TpuConfig, + UpdateClusterRequest, + UpdateMasterRequest, + UpdateNodePoolRequest, + UpgradeAvailableEvent, + UpgradeEvent, + UpgradeResourceType, + UsableSubnetwork, + UsableSubnetworkSecondaryRange, + VerticalPodAutoscaling, + VirtualNIC, + WindowsVersions, + WorkloadCertificates, + WorkloadIdentityConfig, + WorkloadMetadataConfig, +) __all__ = ( "ClusterManagerAsyncClient", diff --git a/google/cloud/container_v1beta1/services/cluster_manager/__init__.py b/google/cloud/container_v1beta1/services/cluster_manager/__init__.py index da06b070..8445ce15 100644 --- a/google/cloud/container_v1beta1/services/cluster_manager/__init__.py +++ b/google/cloud/container_v1beta1/services/cluster_manager/__init__.py @@ -13,8 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # -from .client import ClusterManagerClient from .async_client import ClusterManagerAsyncClient +from .client import ClusterManagerClient __all__ = ( "ClusterManagerClient", diff --git a/google/cloud/container_v1beta1/services/cluster_manager/async_client.py b/google/cloud/container_v1beta1/services/cluster_manager/async_client.py index cf65b96d..212e56e4 100644 --- a/google/cloud/container_v1beta1/services/cluster_manager/async_client.py +++ b/google/cloud/container_v1beta1/services/cluster_manager/async_client.py @@ -16,28 +16,30 @@ from collections import OrderedDict import functools import re -from typing import Dict, Optional, Sequence, Tuple, Type, Union -import pkg_resources +from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union import warnings -from google.api_core.client_options import ClientOptions from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 from google.api_core import retry as retries +from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore +import pkg_resources try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] except AttributeError: # pragma: NO COVER OptionalRetry = Union[retries.Retry, object] # type: ignore +from google.rpc import status_pb2 # type: ignore + from google.cloud.container_v1beta1.services.cluster_manager import pagers from google.cloud.container_v1beta1.types import cluster_service -from google.rpc import status_pb2 # type: ignore -from .transports.base import ClusterManagerTransport, DEFAULT_CLIENT_INFO -from .transports.grpc_asyncio import ClusterManagerGrpcAsyncIOTransport + from .client import ClusterManagerClient +from .transports.base import DEFAULT_CLIENT_INFO, ClusterManagerTransport +from .transports.grpc_asyncio import ClusterManagerGrpcAsyncIOTransport class ClusterManagerAsyncClient: @@ -217,14 +219,13 @@ async def list_clusters( r"""Lists all clusters owned by a project in either the specified zone or all zones. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_list_clusters(): + async def sample_list_clusters(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.ListClustersRequest( @@ -233,7 +234,7 @@ def sample_list_clusters(): ) # Make the request - response = client.list_clusters(request=request) + response = await client.list_clusters(request=request) # Handle the response print(response) @@ -318,7 +319,12 @@ def sample_list_clusters(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -340,9 +346,9 @@ async def get_cluster( from google.cloud import container_v1beta1 - def sample_get_cluster(): + async def sample_get_cluster(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.GetClusterRequest( @@ -352,7 +358,7 @@ def sample_get_cluster(): ) # Make the request - response = client.get_cluster(request=request) + response = await client.get_cluster(request=request) # Handle the response print(response) @@ -446,7 +452,12 @@ def sample_get_cluster(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -476,14 +487,13 @@ async def create_cluster( Finally, an entry is added to the project's global metadata indicating which CIDR range the cluster is using. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_create_cluster(): + async def sample_create_cluster(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.CreateClusterRequest( @@ -492,7 +502,7 @@ def sample_create_cluster(): ) # Make the request - response = client.create_cluster(request=request) + response = await client.create_cluster(request=request) # Handle the response print(response) @@ -578,7 +588,12 @@ def sample_create_cluster(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -601,9 +616,9 @@ async def update_cluster( from google.cloud import container_v1beta1 - def sample_update_cluster(): + async def sample_update_cluster(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.UpdateClusterRequest( @@ -613,7 +628,7 @@ def sample_update_cluster(): ) # Make the request - response = client.update_cluster(request=request) + response = await client.update_cluster(request=request) # Handle the response print(response) @@ -710,7 +725,12 @@ def sample_update_cluster(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -726,14 +746,13 @@ async def update_node_pool( r"""Updates the version and/or image type of a specific node pool. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_update_node_pool(): + async def sample_update_node_pool(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.UpdateNodePoolRequest( @@ -746,7 +765,7 @@ def sample_update_node_pool(): ) # Make the request - response = client.update_node_pool(request=request) + response = await client.update_node_pool(request=request) # Handle the response print(response) @@ -787,7 +806,12 @@ def sample_update_node_pool(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -803,14 +827,13 @@ async def set_node_pool_autoscaling( r"""Sets the autoscaling settings of a specific node pool. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_set_node_pool_autoscaling(): + async def sample_set_node_pool_autoscaling(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetNodePoolAutoscalingRequest( @@ -821,7 +844,7 @@ def sample_set_node_pool_autoscaling(): ) # Make the request - response = client.set_node_pool_autoscaling(request=request) + response = await client.set_node_pool_autoscaling(request=request) # Handle the response print(response) @@ -862,7 +885,12 @@ def sample_set_node_pool_autoscaling(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -885,9 +913,9 @@ async def set_logging_service( from google.cloud import container_v1beta1 - def sample_set_logging_service(): + async def sample_set_logging_service(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetLoggingServiceRequest( @@ -898,7 +926,7 @@ def sample_set_logging_service(): ) # Make the request - response = client.set_logging_service(request=request) + response = await client.set_logging_service(request=request) # Handle the response print(response) @@ -1007,7 +1035,12 @@ def sample_set_logging_service(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1030,9 +1063,9 @@ async def set_monitoring_service( from google.cloud import container_v1beta1 - def sample_set_monitoring_service(): + async def sample_set_monitoring_service(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetMonitoringServiceRequest( @@ -1043,7 +1076,7 @@ def sample_set_monitoring_service(): ) # Make the request - response = client.set_monitoring_service(request=request) + response = await client.set_monitoring_service(request=request) # Handle the response print(response) @@ -1154,7 +1187,12 @@ def sample_set_monitoring_service(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1177,9 +1215,9 @@ async def set_addons_config( from google.cloud import container_v1beta1 - def sample_set_addons_config(): + async def sample_set_addons_config(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetAddonsConfigRequest( @@ -1189,7 +1227,7 @@ def sample_set_addons_config(): ) # Make the request - response = client.set_addons_config(request=request) + response = await client.set_addons_config(request=request) # Handle the response print(response) @@ -1287,7 +1325,12 @@ def sample_set_addons_config(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1308,14 +1351,13 @@ async def set_locations( `projects.locations.clusters.update `__ instead. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_set_locations(): + async def sample_set_locations(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetLocationsRequest( @@ -1326,7 +1368,7 @@ def sample_set_locations(): ) # Make the request - response = client.set_locations(request=request) + response = await client.set_locations(request=request) # Handle the response print(response) @@ -1434,7 +1476,12 @@ def sample_set_locations(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1457,9 +1504,9 @@ async def update_master( from google.cloud import container_v1beta1 - def sample_update_master(): + async def sample_update_master(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.UpdateMasterRequest( @@ -1470,7 +1517,7 @@ def sample_update_master(): ) # Make the request - response = client.update_master(request=request) + response = await client.update_master(request=request) # Handle the response print(response) @@ -1579,7 +1626,12 @@ def sample_update_master(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1597,14 +1649,13 @@ async def set_master_auth( either via password generation or explicitly setting the password. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_set_master_auth(): + async def sample_set_master_auth(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetMasterAuthRequest( @@ -1615,7 +1666,7 @@ def sample_set_master_auth(): ) # Make the request - response = client.set_master_auth(request=request) + response = await client.set_master_auth(request=request) # Handle the response print(response) @@ -1656,7 +1707,12 @@ def sample_set_master_auth(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1683,14 +1739,13 @@ async def delete_cluster( not deleted if they weren't present when the cluster was initially created. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_delete_cluster(): + async def sample_delete_cluster(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.DeleteClusterRequest( @@ -1700,7 +1755,7 @@ def sample_delete_cluster(): ) # Make the request - response = client.delete_cluster(request=request) + response = await client.delete_cluster(request=request) # Handle the response print(response) @@ -1798,7 +1853,12 @@ def sample_delete_cluster(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1816,14 +1876,13 @@ async def list_operations( r"""Lists all operations in a project in the specified zone or all zones. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_list_operations(): + async def sample_list_operations(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.ListOperationsRequest( @@ -1832,7 +1891,7 @@ def sample_list_operations(): ) # Make the request - response = client.list_operations(request=request) + response = await client.list_operations(request=request) # Handle the response print(response) @@ -1918,7 +1977,12 @@ def sample_list_operations(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1940,9 +2004,9 @@ async def get_operation( from google.cloud import container_v1beta1 - def sample_get_operation(): + async def sample_get_operation(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.GetOperationRequest( @@ -1952,7 +2016,7 @@ def sample_get_operation(): ) # Make the request - response = client.get_operation(request=request) + response = await client.get_operation(request=request) # Handle the response print(response) @@ -2049,7 +2113,12 @@ def sample_get_operation(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2071,9 +2140,9 @@ async def cancel_operation( from google.cloud import container_v1beta1 - def sample_cancel_operation(): + async def sample_cancel_operation(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.CancelOperationRequest( @@ -2083,7 +2152,7 @@ def sample_cancel_operation(): ) # Make the request - client.cancel_operation(request=request) + await client.cancel_operation(request=request) Args: request (Union[google.cloud.container_v1beta1.types.CancelOperationRequest, dict]): @@ -2160,7 +2229,10 @@ def sample_cancel_operation(): # Send the request. await rpc( - request, retry=retry, timeout=timeout, metadata=metadata, + request, + retry=retry, + timeout=timeout, + metadata=metadata, ) async def get_server_config( @@ -2176,14 +2248,13 @@ async def get_server_config( r"""Returns configuration info about the Google Kubernetes Engine service. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_get_server_config(): + async def sample_get_server_config(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.GetServerConfigRequest( @@ -2192,7 +2263,7 @@ def sample_get_server_config(): ) # Make the request - response = client.get_server_config(request=request) + response = await client.get_server_config(request=request) # Handle the response print(response) @@ -2277,7 +2348,12 @@ def sample_get_server_config(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2299,9 +2375,9 @@ async def list_node_pools( from google.cloud import container_v1beta1 - def sample_list_node_pools(): + async def sample_list_node_pools(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.ListNodePoolsRequest( @@ -2311,7 +2387,7 @@ def sample_list_node_pools(): ) # Make the request - response = client.list_node_pools(request=request) + response = await client.list_node_pools(request=request) # Handle the response print(response) @@ -2406,7 +2482,12 @@ def sample_list_node_pools(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2424,21 +2505,20 @@ async def get_json_web_keys( This API is not yet intended for general use, and is not available for all clusters. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_get_json_web_keys(): + async def sample_get_json_web_keys(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.GetJSONWebKeysRequest( ) # Make the request - response = client.get_json_web_keys(request=request) + response = await client.get_json_web_keys(request=request) # Handle the response print(response) @@ -2481,7 +2561,12 @@ def sample_get_json_web_keys(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2504,9 +2589,9 @@ async def get_node_pool( from google.cloud import container_v1beta1 - def sample_get_node_pool(): + async def sample_get_node_pool(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.GetNodePoolRequest( @@ -2517,7 +2602,7 @@ def sample_get_node_pool(): ) # Make the request - response = client.get_node_pool(request=request) + response = await client.get_node_pool(request=request) # Handle the response print(response) @@ -2664,7 +2749,12 @@ def sample_get_node_pool(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2687,9 +2777,9 @@ async def create_node_pool( from google.cloud import container_v1beta1 - def sample_create_node_pool(): + async def sample_create_node_pool(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.CreateNodePoolRequest( @@ -2699,7 +2789,7 @@ def sample_create_node_pool(): ) # Make the request - response = client.create_node_pool(request=request) + response = await client.create_node_pool(request=request) # Handle the response print(response) @@ -2793,7 +2883,12 @@ def sample_create_node_pool(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2816,9 +2911,9 @@ async def delete_node_pool( from google.cloud import container_v1beta1 - def sample_delete_node_pool(): + async def sample_delete_node_pool(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.DeleteNodePoolRequest( @@ -2829,7 +2924,7 @@ def sample_delete_node_pool(): ) # Make the request - response = client.delete_node_pool(request=request) + response = await client.delete_node_pool(request=request) # Handle the response print(response) @@ -2937,7 +3032,12 @@ def sample_delete_node_pool(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2958,14 +3058,13 @@ async def rollback_node_pool_upgrade( upgrade. This makes no changes if the last upgrade successfully completed. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_rollback_node_pool_upgrade(): + async def sample_rollback_node_pool_upgrade(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.RollbackNodePoolUpgradeRequest( @@ -2976,7 +3075,7 @@ def sample_rollback_node_pool_upgrade(): ) # Make the request - response = client.rollback_node_pool_upgrade(request=request) + response = await client.rollback_node_pool_upgrade(request=request) # Handle the response print(response) @@ -3077,7 +3176,12 @@ def sample_rollback_node_pool_upgrade(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3101,9 +3205,9 @@ async def set_node_pool_management( from google.cloud import container_v1beta1 - def sample_set_node_pool_management(): + async def sample_set_node_pool_management(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetNodePoolManagementRequest( @@ -3114,7 +3218,7 @@ def sample_set_node_pool_management(): ) # Make the request - response = client.set_node_pool_management(request=request) + response = await client.set_node_pool_management(request=request) # Handle the response print(response) @@ -3224,7 +3328,12 @@ def sample_set_node_pool_management(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3236,9 +3345,7 @@ async def set_labels( project_id: str = None, zone: str = None, cluster_id: str = None, - resource_labels: Sequence[ - cluster_service.SetLabelsRequest.ResourceLabelsEntry - ] = None, + resource_labels: Mapping[str, str] = None, label_fingerprint: str = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, timeout: float = None, @@ -3250,9 +3357,9 @@ async def set_labels( from google.cloud import container_v1beta1 - def sample_set_labels(): + async def sample_set_labels(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetLabelsRequest( @@ -3263,7 +3370,7 @@ def sample_set_labels(): ) # Make the request - response = client.set_labels(request=request) + response = await client.set_labels(request=request) # Handle the response print(response) @@ -3302,7 +3409,7 @@ def sample_set_labels(): This corresponds to the ``cluster_id`` field on the ``request`` instance; if ``request`` is provided, this should not be set. - resource_labels (:class:`Sequence[google.cloud.container_v1beta1.types.SetLabelsRequest.ResourceLabelsEntry]`): + resource_labels (:class:`Mapping[str, str]`): Required. The labels to set for that cluster. @@ -3379,7 +3486,12 @@ def sample_set_labels(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3399,14 +3511,13 @@ async def set_legacy_abac( r"""Enables or disables the ABAC authorization mechanism on a cluster. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_set_legacy_abac(): + async def sample_set_legacy_abac(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetLegacyAbacRequest( @@ -3417,7 +3528,7 @@ def sample_set_legacy_abac(): ) # Make the request - response = client.set_legacy_abac(request=request) + response = await client.set_legacy_abac(request=request) # Handle the response print(response) @@ -3514,7 +3625,12 @@ def sample_set_legacy_abac(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3536,9 +3652,9 @@ async def start_ip_rotation( from google.cloud import container_v1beta1 - def sample_start_ip_rotation(): + async def sample_start_ip_rotation(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.StartIPRotationRequest( @@ -3548,7 +3664,7 @@ def sample_start_ip_rotation(): ) # Make the request - response = client.start_ip_rotation(request=request) + response = await client.start_ip_rotation(request=request) # Handle the response print(response) @@ -3636,7 +3752,12 @@ def sample_start_ip_rotation(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3658,9 +3779,9 @@ async def complete_ip_rotation( from google.cloud import container_v1beta1 - def sample_complete_ip_rotation(): + async def sample_complete_ip_rotation(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.CompleteIPRotationRequest( @@ -3670,7 +3791,7 @@ def sample_complete_ip_rotation(): ) # Make the request - response = client.complete_ip_rotation(request=request) + response = await client.complete_ip_rotation(request=request) # Handle the response print(response) @@ -3757,7 +3878,12 @@ def sample_complete_ip_rotation(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3775,14 +3901,13 @@ async def set_node_pool_size( created by modifying [NodePool.locations][google.container.v1beta1.NodePool.locations]. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_set_node_pool_size(): + async def sample_set_node_pool_size(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetNodePoolSizeRequest( @@ -3794,7 +3919,7 @@ def sample_set_node_pool_size(): ) # Make the request - response = client.set_node_pool_size(request=request) + response = await client.set_node_pool_size(request=request) # Handle the response print(response) @@ -3835,7 +3960,12 @@ def sample_set_node_pool_size(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3858,9 +3988,9 @@ async def set_network_policy( from google.cloud import container_v1beta1 - def sample_set_network_policy(): + async def sample_set_network_policy(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetNetworkPolicyRequest( @@ -3870,7 +4000,7 @@ def sample_set_network_policy(): ) # Make the request - response = client.set_network_policy(request=request) + response = await client.set_network_policy(request=request) # Handle the response print(response) @@ -3966,7 +4096,12 @@ def sample_set_network_policy(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3989,9 +4124,9 @@ async def set_maintenance_policy( from google.cloud import container_v1beta1 - def sample_set_maintenance_policy(): + async def sample_set_maintenance_policy(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.SetMaintenancePolicyRequest( @@ -4001,7 +4136,7 @@ def sample_set_maintenance_policy(): ) # Make the request - response = client.set_maintenance_policy(request=request) + response = await client.set_maintenance_policy(request=request) # Handle the response print(response) @@ -4093,7 +4228,12 @@ def sample_set_maintenance_policy(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -4110,14 +4250,13 @@ async def list_usable_subnetworks( r"""Lists subnetworks that can be used for creating clusters in a project. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_list_usable_subnetworks(): + async def sample_list_usable_subnetworks(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.ListUsableSubnetworksRequest( @@ -4128,7 +4267,7 @@ def sample_list_usable_subnetworks(): page_result = client.list_usable_subnetworks(request=request) # Handle the response - for response in page_result: + async for response in page_result: print(response) Args: @@ -4202,12 +4341,20 @@ def sample_list_usable_subnetworks(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__aiter__` convenience method. response = pagers.ListUsableSubnetworksAsyncPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -4225,14 +4372,13 @@ async def list_locations( r"""Fetches locations that offer Google Kubernetes Engine. - .. code-block:: python from google.cloud import container_v1beta1 - def sample_list_locations(): + async def sample_list_locations(): # Create a client - client = container_v1beta1.ClusterManagerClient() + client = container_v1beta1.ClusterManagerAsyncClient() # Initialize request argument(s) request = container_v1beta1.ListLocationsRequest( @@ -4240,7 +4386,7 @@ def sample_list_locations(): ) # Make the request - response = client.list_locations(request=request) + response = await client.list_locations(request=request) # Handle the response print(response) @@ -4311,7 +4457,12 @@ def sample_list_locations(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -4325,7 +4476,9 @@ async def __aexit__(self, exc_type, exc, tb): try: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution("google-cloud-container",).version, + gapic_version=pkg_resources.get_distribution( + "google-cloud-container", + ).version, ) except pkg_resources.DistributionNotFound: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() diff --git a/google/cloud/container_v1beta1/services/cluster_manager/client.py b/google/cloud/container_v1beta1/services/cluster_manager/client.py index 8de0359d..50360b33 100644 --- a/google/cloud/container_v1beta1/services/cluster_manager/client.py +++ b/google/cloud/container_v1beta1/services/cluster_manager/client.py @@ -16,8 +16,7 @@ from collections import OrderedDict import os import re -from typing import Dict, Optional, Sequence, Tuple, Type, Union -import pkg_resources +from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union import warnings from google.api_core import client_options as client_options_lib @@ -25,20 +24,23 @@ from google.api_core import gapic_v1 from google.api_core import retry as retries from google.auth import credentials as ga_credentials # type: ignore +from google.auth.exceptions import MutualTLSChannelError # type: ignore from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore -from google.auth.exceptions import MutualTLSChannelError # type: ignore from google.oauth2 import service_account # type: ignore +import pkg_resources try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] except AttributeError: # pragma: NO COVER OptionalRetry = Union[retries.Retry, object] # type: ignore +from google.rpc import status_pb2 # type: ignore + from google.cloud.container_v1beta1.services.cluster_manager import pagers from google.cloud.container_v1beta1.types import cluster_service -from google.rpc import status_pb2 # type: ignore -from .transports.base import ClusterManagerTransport, DEFAULT_CLIENT_INFO + +from .transports.base import DEFAULT_CLIENT_INFO, ClusterManagerTransport from .transports.grpc import ClusterManagerGrpcTransport from .transports.grpc_asyncio import ClusterManagerGrpcAsyncIOTransport @@ -57,7 +59,10 @@ class ClusterManagerClientMeta(type): _transport_registry["grpc"] = ClusterManagerGrpcTransport _transport_registry["grpc_asyncio"] = ClusterManagerGrpcAsyncIOTransport - def get_transport_class(cls, label: str = None,) -> Type[ClusterManagerTransport]: + def get_transport_class( + cls, + label: str = None, + ) -> Type[ClusterManagerTransport]: """Returns an appropriate transport class. Args: @@ -162,9 +167,15 @@ def transport(self) -> ClusterManagerTransport: return self._transport @staticmethod - def topic_path(project: str, topic: str,) -> str: + def topic_path( + project: str, + topic: str, + ) -> str: """Returns a fully-qualified topic string.""" - return "projects/{project}/topics/{topic}".format(project=project, topic=topic,) + return "projects/{project}/topics/{topic}".format( + project=project, + topic=topic, + ) @staticmethod def parse_topic_path(path: str) -> Dict[str, str]: @@ -173,7 +184,9 @@ def parse_topic_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_billing_account_path(billing_account: str,) -> str: + def common_billing_account_path( + billing_account: str, + ) -> str: """Returns a fully-qualified billing_account string.""" return "billingAccounts/{billing_account}".format( billing_account=billing_account, @@ -186,9 +199,13 @@ def parse_common_billing_account_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_folder_path(folder: str,) -> str: + def common_folder_path( + folder: str, + ) -> str: """Returns a fully-qualified folder string.""" - return "folders/{folder}".format(folder=folder,) + return "folders/{folder}".format( + folder=folder, + ) @staticmethod def parse_common_folder_path(path: str) -> Dict[str, str]: @@ -197,9 +214,13 @@ def parse_common_folder_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_organization_path(organization: str,) -> str: + def common_organization_path( + organization: str, + ) -> str: """Returns a fully-qualified organization string.""" - return "organizations/{organization}".format(organization=organization,) + return "organizations/{organization}".format( + organization=organization, + ) @staticmethod def parse_common_organization_path(path: str) -> Dict[str, str]: @@ -208,9 +229,13 @@ def parse_common_organization_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_project_path(project: str,) -> str: + def common_project_path( + project: str, + ) -> str: """Returns a fully-qualified project string.""" - return "projects/{project}".format(project=project,) + return "projects/{project}".format( + project=project, + ) @staticmethod def parse_common_project_path(path: str) -> Dict[str, str]: @@ -219,10 +244,14 @@ def parse_common_project_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_location_path(project: str, location: str,) -> str: + def common_location_path( + project: str, + location: str, + ) -> str: """Returns a fully-qualified location string.""" return "projects/{project}/locations/{location}".format( - project=project, location=location, + project=project, + location=location, ) @staticmethod @@ -409,7 +438,6 @@ def list_clusters( r"""Lists all clusters owned by a project in either the specified zone or all zones. - .. code-block:: python from google.cloud import container_v1beta1 @@ -500,7 +528,12 @@ def sample_list_clusters(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -618,7 +651,12 @@ def sample_get_cluster(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -648,7 +686,6 @@ def create_cluster( Finally, an entry is added to the project's global metadata indicating which CIDR range the cluster is using. - .. code-block:: python from google.cloud import container_v1beta1 @@ -750,7 +787,12 @@ def sample_create_cluster(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -882,7 +924,12 @@ def sample_update_cluster(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -898,7 +945,6 @@ def update_node_pool( r"""Updates the version and/or image type of a specific node pool. - .. code-block:: python from google.cloud import container_v1beta1 @@ -960,7 +1006,12 @@ def sample_update_node_pool(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -976,7 +1027,6 @@ def set_node_pool_autoscaling( r"""Sets the autoscaling settings of a specific node pool. - .. code-block:: python from google.cloud import container_v1beta1 @@ -1038,7 +1088,12 @@ def sample_set_node_pool_autoscaling(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1183,7 +1238,12 @@ def sample_set_logging_service(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1330,7 +1390,12 @@ def sample_set_monitoring_service(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1463,7 +1528,12 @@ def sample_set_addons_config(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1484,7 +1554,6 @@ def set_locations( `projects.locations.clusters.update `__ instead. - .. code-block:: python from google.cloud import container_v1beta1 @@ -1610,7 +1679,12 @@ def sample_set_locations(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1755,7 +1829,12 @@ def sample_update_master(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1773,7 +1852,6 @@ def set_master_auth( either via password generation or explicitly setting the password. - .. code-block:: python from google.cloud import container_v1beta1 @@ -1833,7 +1911,12 @@ def sample_set_master_auth(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1860,7 +1943,6 @@ def delete_cluster( not deleted if they weren't present when the cluster was initially created. - .. code-block:: python from google.cloud import container_v1beta1 @@ -1965,7 +2047,12 @@ def sample_delete_cluster(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1983,7 +2070,6 @@ def list_operations( r"""Lists all operations in a project in the specified zone or all zones. - .. code-block:: python from google.cloud import container_v1beta1 @@ -2075,7 +2161,12 @@ def sample_list_operations(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2196,7 +2287,12 @@ def sample_get_operation(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2307,7 +2403,10 @@ def sample_cancel_operation(): # Send the request. rpc( - request, retry=retry, timeout=timeout, metadata=metadata, + request, + retry=retry, + timeout=timeout, + metadata=metadata, ) def get_server_config( @@ -2323,7 +2422,6 @@ def get_server_config( r"""Returns configuration info about the Google Kubernetes Engine service. - .. code-block:: python from google.cloud import container_v1beta1 @@ -2414,7 +2512,12 @@ def sample_get_server_config(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2533,7 +2636,12 @@ def sample_list_node_pools(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2551,7 +2659,6 @@ def get_json_web_keys( This API is not yet intended for general use, and is not available for all clusters. - .. code-block:: python from google.cloud import container_v1beta1 @@ -2609,7 +2716,12 @@ def sample_get_json_web_keys(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2782,7 +2894,12 @@ def sample_get_node_pool(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -2911,7 +3028,12 @@ def sample_create_node_pool(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3045,7 +3167,12 @@ def sample_delete_node_pool(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3066,7 +3193,6 @@ def rollback_node_pool_upgrade( upgrade. This makes no changes if the last upgrade successfully completed. - .. code-block:: python from google.cloud import container_v1beta1 @@ -3187,7 +3313,12 @@ def sample_rollback_node_pool_upgrade(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3334,7 +3465,12 @@ def sample_set_node_pool_management(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3346,9 +3482,7 @@ def set_labels( project_id: str = None, zone: str = None, cluster_id: str = None, - resource_labels: Sequence[ - cluster_service.SetLabelsRequest.ResourceLabelsEntry - ] = None, + resource_labels: Mapping[str, str] = None, label_fingerprint: str = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, timeout: float = None, @@ -3412,7 +3546,7 @@ def sample_set_labels(): This corresponds to the ``cluster_id`` field on the ``request`` instance; if ``request`` is provided, this should not be set. - resource_labels (Sequence[google.cloud.container_v1beta1.types.SetLabelsRequest.ResourceLabelsEntry]): + resource_labels (Mapping[str, str]): Required. The labels to set for that cluster. @@ -3488,7 +3622,12 @@ def sample_set_labels(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3508,7 +3647,6 @@ def set_legacy_abac( r"""Enables or disables the ABAC authorization mechanism on a cluster. - .. code-block:: python from google.cloud import container_v1beta1 @@ -3623,7 +3761,12 @@ def sample_set_legacy_abac(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3745,7 +3888,12 @@ def sample_start_ip_rotation(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3866,7 +4014,12 @@ def sample_complete_ip_rotation(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -3884,7 +4037,6 @@ def set_node_pool_size( created by modifying [NodePool.locations][google.container.v1beta1.NodePool.locations]. - .. code-block:: python from google.cloud import container_v1beta1 @@ -3945,7 +4097,12 @@ def sample_set_node_pool_size(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -4076,7 +4233,12 @@ def sample_set_network_policy(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -4203,7 +4365,12 @@ def sample_set_maintenance_policy(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -4220,7 +4387,6 @@ def list_usable_subnetworks( r"""Lists subnetworks that can be used for creating clusters in a project. - .. code-block:: python from google.cloud import container_v1beta1 @@ -4302,12 +4468,20 @@ def sample_list_usable_subnetworks(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__iter__` convenience method. response = pagers.ListUsableSubnetworksPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -4325,7 +4499,6 @@ def list_locations( r"""Fetches locations that offer Google Kubernetes Engine. - .. code-block:: python from google.cloud import container_v1beta1 @@ -4401,7 +4574,12 @@ def sample_list_locations(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -4422,7 +4600,9 @@ def __exit__(self, type, value, traceback): try: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution("google-cloud-container",).version, + gapic_version=pkg_resources.get_distribution( + "google-cloud-container", + ).version, ) except pkg_resources.DistributionNotFound: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() diff --git a/google/cloud/container_v1beta1/services/cluster_manager/pagers.py b/google/cloud/container_v1beta1/services/cluster_manager/pagers.py index a92dc3ba..44036d87 100644 --- a/google/cloud/container_v1beta1/services/cluster_manager/pagers.py +++ b/google/cloud/container_v1beta1/services/cluster_manager/pagers.py @@ -18,10 +18,10 @@ AsyncIterator, Awaitable, Callable, + Iterator, + Optional, Sequence, Tuple, - Optional, - Iterator, ) from google.cloud.container_v1beta1.types import cluster_service diff --git a/google/cloud/container_v1beta1/services/cluster_manager/transports/__init__.py b/google/cloud/container_v1beta1/services/cluster_manager/transports/__init__.py index b6375e92..b794d72c 100644 --- a/google/cloud/container_v1beta1/services/cluster_manager/transports/__init__.py +++ b/google/cloud/container_v1beta1/services/cluster_manager/transports/__init__.py @@ -20,7 +20,6 @@ from .grpc import ClusterManagerGrpcTransport from .grpc_asyncio import ClusterManagerGrpcAsyncIOTransport - # Compile a registry of transports. _transport_registry = OrderedDict() # type: Dict[str, Type[ClusterManagerTransport]] _transport_registry["grpc"] = ClusterManagerGrpcTransport diff --git a/google/cloud/container_v1beta1/services/cluster_manager/transports/base.py b/google/cloud/container_v1beta1/services/cluster_manager/transports/base.py index 4cdb5b05..2d54204d 100644 --- a/google/cloud/container_v1beta1/services/cluster_manager/transports/base.py +++ b/google/cloud/container_v1beta1/services/cluster_manager/transports/base.py @@ -15,22 +15,24 @@ # import abc from typing import Awaitable, Callable, Dict, Optional, Sequence, Union -import pkg_resources -import google.auth # type: ignore import google.api_core from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 from google.api_core import retry as retries +import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore +from google.protobuf import empty_pb2 # type: ignore +import pkg_resources from google.cloud.container_v1beta1.types import cluster_service -from google.protobuf import empty_pb2 # type: ignore try: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( - gapic_version=pkg_resources.get_distribution("google-cloud-container",).version, + gapic_version=pkg_resources.get_distribution( + "google-cloud-container", + ).version, ) except pkg_resources.DistributionNotFound: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() @@ -79,6 +81,7 @@ def __init__( always_use_jwt_access (Optional[bool]): Whether self signed JWT should be used for service account credentials. """ + # Save the hostname. Default to port 443 (HTTPS) if none is specified. if ":" not in host: host += ":443" @@ -150,13 +153,19 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.create_cluster: gapic_v1.method.wrap_method( - self.create_cluster, default_timeout=45.0, client_info=client_info, + self.create_cluster, + default_timeout=45.0, + client_info=client_info, ), self.update_cluster: gapic_v1.method.wrap_method( - self.update_cluster, default_timeout=45.0, client_info=client_info, + self.update_cluster, + default_timeout=45.0, + client_info=client_info, ), self.update_node_pool: gapic_v1.method.wrap_method( - self.update_node_pool, default_timeout=45.0, client_info=client_info, + self.update_node_pool, + default_timeout=45.0, + client_info=client_info, ), self.set_node_pool_autoscaling: gapic_v1.method.wrap_method( self.set_node_pool_autoscaling, @@ -164,7 +173,9 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.set_logging_service: gapic_v1.method.wrap_method( - self.set_logging_service, default_timeout=45.0, client_info=client_info, + self.set_logging_service, + default_timeout=45.0, + client_info=client_info, ), self.set_monitoring_service: gapic_v1.method.wrap_method( self.set_monitoring_service, @@ -172,16 +183,24 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.set_addons_config: gapic_v1.method.wrap_method( - self.set_addons_config, default_timeout=45.0, client_info=client_info, + self.set_addons_config, + default_timeout=45.0, + client_info=client_info, ), self.set_locations: gapic_v1.method.wrap_method( - self.set_locations, default_timeout=45.0, client_info=client_info, + self.set_locations, + default_timeout=45.0, + client_info=client_info, ), self.update_master: gapic_v1.method.wrap_method( - self.update_master, default_timeout=45.0, client_info=client_info, + self.update_master, + default_timeout=45.0, + client_info=client_info, ), self.set_master_auth: gapic_v1.method.wrap_method( - self.set_master_auth, default_timeout=45.0, client_info=client_info, + self.set_master_auth, + default_timeout=45.0, + client_info=client_info, ), self.delete_cluster: gapic_v1.method.wrap_method( self.delete_cluster, @@ -229,7 +248,9 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.cancel_operation: gapic_v1.method.wrap_method( - self.cancel_operation, default_timeout=45.0, client_info=client_info, + self.cancel_operation, + default_timeout=45.0, + client_info=client_info, ), self.get_server_config: gapic_v1.method.wrap_method( self.get_server_config, @@ -262,7 +283,9 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.get_json_web_keys: gapic_v1.method.wrap_method( - self.get_json_web_keys, default_timeout=None, client_info=client_info, + self.get_json_web_keys, + default_timeout=None, + client_info=client_info, ), self.get_node_pool: gapic_v1.method.wrap_method( self.get_node_pool, @@ -280,7 +303,9 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.create_node_pool: gapic_v1.method.wrap_method( - self.create_node_pool, default_timeout=45.0, client_info=client_info, + self.create_node_pool, + default_timeout=45.0, + client_info=client_info, ), self.delete_node_pool: gapic_v1.method.wrap_method( self.delete_node_pool, @@ -308,13 +333,19 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.set_labels: gapic_v1.method.wrap_method( - self.set_labels, default_timeout=45.0, client_info=client_info, + self.set_labels, + default_timeout=45.0, + client_info=client_info, ), self.set_legacy_abac: gapic_v1.method.wrap_method( - self.set_legacy_abac, default_timeout=45.0, client_info=client_info, + self.set_legacy_abac, + default_timeout=45.0, + client_info=client_info, ), self.start_ip_rotation: gapic_v1.method.wrap_method( - self.start_ip_rotation, default_timeout=45.0, client_info=client_info, + self.start_ip_rotation, + default_timeout=45.0, + client_info=client_info, ), self.complete_ip_rotation: gapic_v1.method.wrap_method( self.complete_ip_rotation, @@ -322,10 +353,14 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.set_node_pool_size: gapic_v1.method.wrap_method( - self.set_node_pool_size, default_timeout=45.0, client_info=client_info, + self.set_node_pool_size, + default_timeout=45.0, + client_info=client_info, ), self.set_network_policy: gapic_v1.method.wrap_method( - self.set_network_policy, default_timeout=45.0, client_info=client_info, + self.set_network_policy, + default_timeout=45.0, + client_info=client_info, ), self.set_maintenance_policy: gapic_v1.method.wrap_method( self.set_maintenance_policy, @@ -367,9 +402,9 @@ def _prep_wrapped_messages(self, client_info): def close(self): """Closes resources associated with the transport. - .. warning:: - Only call this method if the transport is NOT shared - with other clients - this may cause errors in other clients! + .. warning:: + Only call this method if the transport is NOT shared + with other clients - this may cause errors in other clients! """ raise NotImplementedError() @@ -688,5 +723,9 @@ def list_locations( ]: raise NotImplementedError() + @property + def kind(self) -> str: + raise NotImplementedError() + __all__ = ("ClusterManagerTransport",) diff --git a/google/cloud/container_v1beta1/services/cluster_manager/transports/grpc.py b/google/cloud/container_v1beta1/services/cluster_manager/transports/grpc.py index 358212f7..311b145b 100644 --- a/google/cloud/container_v1beta1/services/cluster_manager/transports/grpc.py +++ b/google/cloud/container_v1beta1/services/cluster_manager/transports/grpc.py @@ -13,20 +13,19 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import warnings from typing import Callable, Dict, Optional, Sequence, Tuple, Union +import warnings -from google.api_core import grpc_helpers -from google.api_core import gapic_v1 +from google.api_core import gapic_v1, grpc_helpers import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore - +from google.protobuf import empty_pb2 # type: ignore import grpc # type: ignore from google.cloud.container_v1beta1.types import cluster_service -from google.protobuf import empty_pb2 # type: ignore -from .base import ClusterManagerTransport, DEFAULT_CLIENT_INFO + +from .base import DEFAULT_CLIENT_INFO, ClusterManagerTransport class ClusterManagerGrpcTransport(ClusterManagerTransport): @@ -225,8 +224,7 @@ def create_channel( @property def grpc_channel(self) -> grpc.Channel: - """Return the channel designed to connect to this service. - """ + """Return the channel designed to connect to this service.""" return self._grpc_channel @property @@ -1161,5 +1159,9 @@ def list_locations( def close(self): self.grpc_channel.close() + @property + def kind(self) -> str: + return "grpc" + __all__ = ("ClusterManagerGrpcTransport",) diff --git a/google/cloud/container_v1beta1/services/cluster_manager/transports/grpc_asyncio.py b/google/cloud/container_v1beta1/services/cluster_manager/transports/grpc_asyncio.py index 9ffacb93..38bc7c68 100644 --- a/google/cloud/container_v1beta1/services/cluster_manager/transports/grpc_asyncio.py +++ b/google/cloud/container_v1beta1/services/cluster_manager/transports/grpc_asyncio.py @@ -13,20 +13,19 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import warnings from typing import Awaitable, Callable, Dict, Optional, Sequence, Tuple, Union +import warnings -from google.api_core import gapic_v1 -from google.api_core import grpc_helpers_async +from google.api_core import gapic_v1, grpc_helpers_async from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore - +from google.protobuf import empty_pb2 # type: ignore import grpc # type: ignore from grpc.experimental import aio # type: ignore from google.cloud.container_v1beta1.types import cluster_service -from google.protobuf import empty_pb2 # type: ignore -from .base import ClusterManagerTransport, DEFAULT_CLIENT_INFO + +from .base import DEFAULT_CLIENT_INFO, ClusterManagerTransport from .grpc import ClusterManagerGrpcTransport diff --git a/google/cloud/container_v1beta1/types/__init__.py b/google/cloud/container_v1beta1/types/__init__.py index 085984e1..c477bff1 100644 --- a/google/cloud/container_v1beta1/types/__init__.py +++ b/google/cloud/container_v1beta1/types/__init__.py @@ -36,6 +36,7 @@ CreateNodePoolRequest, DailyMaintenanceWindow, DatabaseEncryption, + DatapathProvider, DefaultSnatStatus, DeleteClusterRequest, DeleteNodePoolRequest, @@ -111,6 +112,7 @@ PodSecurityPolicyConfig, PrivateClusterConfig, PrivateClusterMasterGlobalAccessConfig, + PrivateIPv6GoogleAccess, RecurringTimeWindow, ReleaseChannel, ReservationAffinity, @@ -144,6 +146,7 @@ UpdateNodePoolRequest, UpgradeAvailableEvent, UpgradeEvent, + UpgradeResourceType, UsableSubnetwork, UsableSubnetworkSecondaryRange, VerticalPodAutoscaling, @@ -152,9 +155,6 @@ WorkloadCertificates, WorkloadIdentityConfig, WorkloadMetadataConfig, - DatapathProvider, - PrivateIPv6GoogleAccess, - UpgradeResourceType, ) __all__ = ( diff --git a/google/cloud/container_v1beta1/types/cluster_service.py b/google/cloud/container_v1beta1/types/cluster_service.py index f5a89bc7..542f625c 100644 --- a/google/cloud/container_v1beta1/types/cluster_service.py +++ b/google/cloud/container_v1beta1/types/cluster_service.py @@ -13,14 +13,12 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import proto # type: ignore - from google.protobuf import timestamp_pb2 # type: ignore from google.protobuf import wrappers_pb2 # type: ignore from google.rpc import code_pb2 # type: ignore from google.rpc import status_pb2 # type: ignore from google.type import date_pb2 # type: ignore - +import proto # type: ignore __protobuf__ = proto.module( package="google.container.v1beta1", @@ -203,7 +201,7 @@ class LinuxNodeConfig(proto.Message): r"""Parameters that can be configured on Linux nodes. Attributes: - sysctls (Sequence[google.cloud.container_v1beta1.types.LinuxNodeConfig.SysctlsEntry]): + sysctls (Mapping[str, str]): The Linux kernel parameters to be applied to the nodes and all pods running on the nodes. @@ -215,7 +213,11 @@ class LinuxNodeConfig(proto.Message): net.ipv4.tcp_tw_reuse """ - sysctls = proto.MapField(proto.STRING, proto.STRING, number=1,) + sysctls = proto.MapField( + proto.STRING, + proto.STRING, + number=1, + ) class NodeKubeletConfig(proto.Message): @@ -257,11 +259,19 @@ class NodeKubeletConfig(proto.Message): The value must be a positive duration. """ - cpu_manager_policy = proto.Field(proto.STRING, number=1,) + cpu_manager_policy = proto.Field( + proto.STRING, + number=1, + ) cpu_cfs_quota = proto.Field( - proto.MESSAGE, number=2, message=wrappers_pb2.BoolValue, + proto.MESSAGE, + number=2, + message=wrappers_pb2.BoolValue, + ) + cpu_cfs_quota_period = proto.Field( + proto.STRING, + number=3, ) - cpu_cfs_quota_period = proto.Field(proto.STRING, number=3,) class NodeConfig(proto.Message): @@ -301,7 +311,7 @@ class NodeConfig(proto.Message): address of the Service Account; otherwise, if no Service Account is specified, the "default" service account is used. - metadata (Sequence[google.cloud.container_v1beta1.types.NodeConfig.MetadataEntry]): + metadata (Mapping[str, str]): The metadata key/value pairs assigned to instances in the cluster. @@ -342,7 +352,7 @@ class NodeConfig(proto.Message): The image type to use for this node. Note that for a given image type, the latest version of it will be used. - labels (Sequence[google.cloud.container_v1beta1.types.NodeConfig.LabelsEntry]): + labels (Mapping[str, str]): The map of Kubernetes labels (key/value pairs) to be applied to each node. These will added in addition to any default label(s) that @@ -445,47 +455,128 @@ class NodeConfig(proto.Message): rebrand of the existing preemptible flag. """ - machine_type = proto.Field(proto.STRING, number=1,) - disk_size_gb = proto.Field(proto.INT32, number=2,) - oauth_scopes = proto.RepeatedField(proto.STRING, number=3,) - service_account = proto.Field(proto.STRING, number=9,) - metadata = proto.MapField(proto.STRING, proto.STRING, number=4,) - image_type = proto.Field(proto.STRING, number=5,) - labels = proto.MapField(proto.STRING, proto.STRING, number=6,) - local_ssd_count = proto.Field(proto.INT32, number=7,) - tags = proto.RepeatedField(proto.STRING, number=8,) - preemptible = proto.Field(proto.BOOL, number=10,) + machine_type = proto.Field( + proto.STRING, + number=1, + ) + disk_size_gb = proto.Field( + proto.INT32, + number=2, + ) + oauth_scopes = proto.RepeatedField( + proto.STRING, + number=3, + ) + service_account = proto.Field( + proto.STRING, + number=9, + ) + metadata = proto.MapField( + proto.STRING, + proto.STRING, + number=4, + ) + image_type = proto.Field( + proto.STRING, + number=5, + ) + labels = proto.MapField( + proto.STRING, + proto.STRING, + number=6, + ) + local_ssd_count = proto.Field( + proto.INT32, + number=7, + ) + tags = proto.RepeatedField( + proto.STRING, + number=8, + ) + preemptible = proto.Field( + proto.BOOL, + number=10, + ) accelerators = proto.RepeatedField( - proto.MESSAGE, number=11, message="AcceleratorConfig", + proto.MESSAGE, + number=11, + message="AcceleratorConfig", + ) + sandbox_config = proto.Field( + proto.MESSAGE, + number=17, + message="SandboxConfig", + ) + node_group = proto.Field( + proto.STRING, + number=18, ) - sandbox_config = proto.Field(proto.MESSAGE, number=17, message="SandboxConfig",) - node_group = proto.Field(proto.STRING, number=18,) reservation_affinity = proto.Field( - proto.MESSAGE, number=19, message="ReservationAffinity", + proto.MESSAGE, + number=19, + message="ReservationAffinity", + ) + disk_type = proto.Field( + proto.STRING, + number=12, + ) + min_cpu_platform = proto.Field( + proto.STRING, + number=13, ) - disk_type = proto.Field(proto.STRING, number=12,) - min_cpu_platform = proto.Field(proto.STRING, number=13,) workload_metadata_config = proto.Field( - proto.MESSAGE, number=14, message="WorkloadMetadataConfig", + proto.MESSAGE, + number=14, + message="WorkloadMetadataConfig", + ) + taints = proto.RepeatedField( + proto.MESSAGE, + number=15, + message="NodeTaint", + ) + boot_disk_kms_key = proto.Field( + proto.STRING, + number=23, ) - taints = proto.RepeatedField(proto.MESSAGE, number=15, message="NodeTaint",) - boot_disk_kms_key = proto.Field(proto.STRING, number=23,) shielded_instance_config = proto.Field( - proto.MESSAGE, number=20, message="ShieldedInstanceConfig", + proto.MESSAGE, + number=20, + message="ShieldedInstanceConfig", ) linux_node_config = proto.Field( - proto.MESSAGE, number=21, message="LinuxNodeConfig", + proto.MESSAGE, + number=21, + message="LinuxNodeConfig", + ) + kubelet_config = proto.Field( + proto.MESSAGE, + number=22, + message="NodeKubeletConfig", ) - kubelet_config = proto.Field(proto.MESSAGE, number=22, message="NodeKubeletConfig",) ephemeral_storage_config = proto.Field( - proto.MESSAGE, number=24, message="EphemeralStorageConfig", + proto.MESSAGE, + number=24, + message="EphemeralStorageConfig", + ) + gcfs_config = proto.Field( + proto.MESSAGE, + number=25, + message="GcfsConfig", ) - gcfs_config = proto.Field(proto.MESSAGE, number=25, message="GcfsConfig",) advanced_machine_features = proto.Field( - proto.MESSAGE, number=26, message="AdvancedMachineFeatures", + proto.MESSAGE, + number=26, + message="AdvancedMachineFeatures", + ) + gvnic = proto.Field( + proto.MESSAGE, + number=29, + message="VirtualNIC", + ) + spot = proto.Field( + proto.BOOL, + number=32, ) - gvnic = proto.Field(proto.MESSAGE, number=29, message="VirtualNIC",) - spot = proto.Field(proto.BOOL, number=32,) class AdvancedMachineFeatures(proto.Message): @@ -502,7 +593,11 @@ class AdvancedMachineFeatures(proto.Message): This field is a member of `oneof`_ ``_threads_per_core``. """ - threads_per_core = proto.Field(proto.INT64, number=1, optional=True,) + threads_per_core = proto.Field( + proto.INT64, + number=1, + optional=True, + ) class NodeNetworkConfig(proto.Message): @@ -556,9 +651,18 @@ class NodeNetworkConfig(proto.Message): created. """ - create_pod_range = proto.Field(proto.BOOL, number=4,) - pod_range = proto.Field(proto.STRING, number=5,) - pod_ipv4_cidr_block = proto.Field(proto.STRING, number=6,) + create_pod_range = proto.Field( + proto.BOOL, + number=4, + ) + pod_range = proto.Field( + proto.STRING, + number=5, + ) + pod_ipv4_cidr_block = proto.Field( + proto.STRING, + number=6, + ) class ShieldedInstanceConfig(proto.Message): @@ -584,8 +688,14 @@ class ShieldedInstanceConfig(proto.Message): is created. """ - enable_secure_boot = proto.Field(proto.BOOL, number=1,) - enable_integrity_monitoring = proto.Field(proto.BOOL, number=2,) + enable_secure_boot = proto.Field( + proto.BOOL, + number=1, + ) + enable_integrity_monitoring = proto.Field( + proto.BOOL, + number=2, + ) class SandboxConfig(proto.Message): @@ -605,8 +715,15 @@ class Type(proto.Enum): UNSPECIFIED = 0 GVISOR = 1 - sandbox_type = proto.Field(proto.STRING, number=1,) - type_ = proto.Field(proto.ENUM, number=2, enum=Type,) + sandbox_type = proto.Field( + proto.STRING, + number=1, + ) + type_ = proto.Field( + proto.ENUM, + number=2, + enum=Type, + ) class EphemeralStorageConfig(proto.Message): @@ -621,7 +738,10 @@ class EphemeralStorageConfig(proto.Message): using local SSDs as ephemeral storage. """ - local_ssd_count = proto.Field(proto.INT32, number=1,) + local_ssd_count = proto.Field( + proto.INT32, + number=1, + ) class GcfsConfig(proto.Message): @@ -633,7 +753,10 @@ class GcfsConfig(proto.Message): Whether to use GCFS. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class ReservationAffinity(proto.Message): @@ -664,9 +787,19 @@ class Type(proto.Enum): ANY_RESERVATION = 2 SPECIFIC_RESERVATION = 3 - consume_reservation_type = proto.Field(proto.ENUM, number=1, enum=Type,) - key = proto.Field(proto.STRING, number=2,) - values = proto.RepeatedField(proto.STRING, number=3,) + consume_reservation_type = proto.Field( + proto.ENUM, + number=1, + enum=Type, + ) + key = proto.Field( + proto.STRING, + number=2, + ) + values = proto.RepeatedField( + proto.STRING, + number=3, + ) class NodeTaint(proto.Message): @@ -694,9 +827,19 @@ class Effect(proto.Enum): PREFER_NO_SCHEDULE = 2 NO_EXECUTE = 3 - key = proto.Field(proto.STRING, number=1,) - value = proto.Field(proto.STRING, number=2,) - effect = proto.Field(proto.ENUM, number=3, enum=Effect,) + key = proto.Field( + proto.STRING, + number=1, + ) + value = proto.Field( + proto.STRING, + number=2, + ) + effect = proto.Field( + proto.ENUM, + number=3, + enum=Effect, + ) class NodeTaints(proto.Message): @@ -708,7 +851,11 @@ class NodeTaints(proto.Message): List of node taints. """ - taints = proto.RepeatedField(proto.MESSAGE, number=1, message="NodeTaint",) + taints = proto.RepeatedField( + proto.MESSAGE, + number=1, + message="NodeTaint", + ) class NodeLabels(proto.Message): @@ -716,11 +863,15 @@ class NodeLabels(proto.Message): labels `__. Attributes: - labels (Sequence[google.cloud.container_v1beta1.types.NodeLabels.LabelsEntry]): + labels (Mapping[str, str]): Map of node label keys and node label values. """ - labels = proto.MapField(proto.STRING, proto.STRING, number=1,) + labels = proto.MapField( + proto.STRING, + proto.STRING, + number=1, + ) class NetworkTags(proto.Message): @@ -733,7 +884,10 @@ class NetworkTags(proto.Message): List of network tags. """ - tags = proto.RepeatedField(proto.STRING, number=1,) + tags = proto.RepeatedField( + proto.STRING, + number=1, + ) class MasterAuth(proto.Message): @@ -781,14 +935,31 @@ class MasterAuth(proto.Message): authenticate to the cluster endpoint. """ - username = proto.Field(proto.STRING, number=1,) - password = proto.Field(proto.STRING, number=2,) + username = proto.Field( + proto.STRING, + number=1, + ) + password = proto.Field( + proto.STRING, + number=2, + ) client_certificate_config = proto.Field( - proto.MESSAGE, number=3, message="ClientCertificateConfig", + proto.MESSAGE, + number=3, + message="ClientCertificateConfig", + ) + cluster_ca_certificate = proto.Field( + proto.STRING, + number=100, + ) + client_certificate = proto.Field( + proto.STRING, + number=101, + ) + client_key = proto.Field( + proto.STRING, + number=102, ) - cluster_ca_certificate = proto.Field(proto.STRING, number=100,) - client_certificate = proto.Field(proto.STRING, number=101,) - client_key = proto.Field(proto.STRING, number=102,) class ClientCertificateConfig(proto.Message): @@ -799,7 +970,10 @@ class ClientCertificateConfig(proto.Message): Issue a client certificate. """ - issue_client_certificate = proto.Field(proto.BOOL, number=1,) + issue_client_certificate = proto.Field( + proto.BOOL, + number=1, + ) class AddonsConfig(proto.Message): @@ -860,32 +1034,64 @@ class AddonsConfig(proto.Message): """ http_load_balancing = proto.Field( - proto.MESSAGE, number=1, message="HttpLoadBalancing", + proto.MESSAGE, + number=1, + message="HttpLoadBalancing", ) horizontal_pod_autoscaling = proto.Field( - proto.MESSAGE, number=2, message="HorizontalPodAutoscaling", + proto.MESSAGE, + number=2, + message="HorizontalPodAutoscaling", ) kubernetes_dashboard = proto.Field( - proto.MESSAGE, number=3, message="KubernetesDashboard", + proto.MESSAGE, + number=3, + message="KubernetesDashboard", ) network_policy_config = proto.Field( - proto.MESSAGE, number=4, message="NetworkPolicyConfig", + proto.MESSAGE, + number=4, + message="NetworkPolicyConfig", + ) + istio_config = proto.Field( + proto.MESSAGE, + number=5, + message="IstioConfig", + ) + cloud_run_config = proto.Field( + proto.MESSAGE, + number=7, + message="CloudRunConfig", + ) + dns_cache_config = proto.Field( + proto.MESSAGE, + number=8, + message="DnsCacheConfig", ) - istio_config = proto.Field(proto.MESSAGE, number=5, message="IstioConfig",) - cloud_run_config = proto.Field(proto.MESSAGE, number=7, message="CloudRunConfig",) - dns_cache_config = proto.Field(proto.MESSAGE, number=8, message="DnsCacheConfig",) config_connector_config = proto.Field( - proto.MESSAGE, number=10, message="ConfigConnectorConfig", + proto.MESSAGE, + number=10, + message="ConfigConnectorConfig", ) gce_persistent_disk_csi_driver_config = proto.Field( - proto.MESSAGE, number=11, message="GcePersistentDiskCsiDriverConfig", + proto.MESSAGE, + number=11, + message="GcePersistentDiskCsiDriverConfig", + ) + kalm_config = proto.Field( + proto.MESSAGE, + number=12, + message="KalmConfig", ) - kalm_config = proto.Field(proto.MESSAGE, number=12, message="KalmConfig",) gcp_filestore_csi_driver_config = proto.Field( - proto.MESSAGE, number=14, message="GcpFilestoreCsiDriverConfig", + proto.MESSAGE, + number=14, + message="GcpFilestoreCsiDriverConfig", ) gke_backup_agent_config = proto.Field( - proto.MESSAGE, number=16, message="GkeBackupAgentConfig", + proto.MESSAGE, + number=16, + message="GkeBackupAgentConfig", ) @@ -902,7 +1108,10 @@ class HttpLoadBalancing(proto.Message): balancers. """ - disabled = proto.Field(proto.BOOL, number=1,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) class HorizontalPodAutoscaling(proto.Message): @@ -919,7 +1128,10 @@ class HorizontalPodAutoscaling(proto.Message): Stackdriver Monitoring. """ - disabled = proto.Field(proto.BOOL, number=1,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) class KubernetesDashboard(proto.Message): @@ -931,7 +1143,10 @@ class KubernetesDashboard(proto.Message): for this cluster. """ - disabled = proto.Field(proto.BOOL, number=1,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) class NetworkPolicyConfig(proto.Message): @@ -945,7 +1160,10 @@ class NetworkPolicyConfig(proto.Message): cluster. """ - disabled = proto.Field(proto.BOOL, number=1,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) class DnsCacheConfig(proto.Message): @@ -957,7 +1175,10 @@ class DnsCacheConfig(proto.Message): this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class KalmConfig(proto.Message): @@ -968,7 +1189,10 @@ class KalmConfig(proto.Message): Whether KALM is enabled for this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class GkeBackupAgentConfig(proto.Message): @@ -980,7 +1204,10 @@ class GkeBackupAgentConfig(proto.Message): for this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class ConfigConnectorConfig(proto.Message): @@ -992,7 +1219,10 @@ class ConfigConnectorConfig(proto.Message): cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class GcePersistentDiskCsiDriverConfig(proto.Message): @@ -1004,7 +1234,10 @@ class GcePersistentDiskCsiDriverConfig(proto.Message): enabled for this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class GcpFilestoreCsiDriverConfig(proto.Message): @@ -1016,7 +1249,10 @@ class GcpFilestoreCsiDriverConfig(proto.Message): enabled for this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class PrivateClusterMasterGlobalAccessConfig(proto.Message): @@ -1028,7 +1264,10 @@ class PrivateClusterMasterGlobalAccessConfig(proto.Message): not. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class PrivateClusterConfig(proto.Message): @@ -1063,14 +1302,34 @@ class PrivateClusterConfig(proto.Message): Controls master global access settings. """ - enable_private_nodes = proto.Field(proto.BOOL, number=1,) - enable_private_endpoint = proto.Field(proto.BOOL, number=2,) - master_ipv4_cidr_block = proto.Field(proto.STRING, number=3,) - private_endpoint = proto.Field(proto.STRING, number=4,) - public_endpoint = proto.Field(proto.STRING, number=5,) - peering_name = proto.Field(proto.STRING, number=7,) + enable_private_nodes = proto.Field( + proto.BOOL, + number=1, + ) + enable_private_endpoint = proto.Field( + proto.BOOL, + number=2, + ) + master_ipv4_cidr_block = proto.Field( + proto.STRING, + number=3, + ) + private_endpoint = proto.Field( + proto.STRING, + number=4, + ) + public_endpoint = proto.Field( + proto.STRING, + number=5, + ) + peering_name = proto.Field( + proto.STRING, + number=7, + ) master_global_access_config = proto.Field( - proto.MESSAGE, number=8, message="PrivateClusterMasterGlobalAccessConfig", + proto.MESSAGE, + number=8, + message="PrivateClusterMasterGlobalAccessConfig", ) @@ -1092,8 +1351,15 @@ class IstioAuthMode(proto.Enum): AUTH_NONE = 0 AUTH_MUTUAL_TLS = 1 - disabled = proto.Field(proto.BOOL, number=1,) - auth = proto.Field(proto.ENUM, number=2, enum=IstioAuthMode,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) + auth = proto.Field( + proto.ENUM, + number=2, + enum=IstioAuthMode, + ) class CloudRunConfig(proto.Message): @@ -1114,8 +1380,15 @@ class LoadBalancerType(proto.Enum): LOAD_BALANCER_TYPE_EXTERNAL = 1 LOAD_BALANCER_TYPE_INTERNAL = 2 - disabled = proto.Field(proto.BOOL, number=1,) - load_balancer_type = proto.Field(proto.ENUM, number=3, enum=LoadBalancerType,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) + load_balancer_type = proto.Field( + proto.ENUM, + number=3, + enum=LoadBalancerType, + ) class MasterAuthorizedNetworksConfig(proto.Message): @@ -1145,11 +1418,24 @@ class CidrBlock(proto.Message): cidr_block must be specified in CIDR notation. """ - display_name = proto.Field(proto.STRING, number=1,) - cidr_block = proto.Field(proto.STRING, number=2,) + display_name = proto.Field( + proto.STRING, + number=1, + ) + cidr_block = proto.Field( + proto.STRING, + number=2, + ) - enabled = proto.Field(proto.BOOL, number=1,) - cidr_blocks = proto.RepeatedField(proto.MESSAGE, number=2, message=CidrBlock,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) + cidr_blocks = proto.RepeatedField( + proto.MESSAGE, + number=2, + message=CidrBlock, + ) class LegacyAbac(proto.Message): @@ -1166,7 +1452,10 @@ class LegacyAbac(proto.Message): configuration or IAM. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class NetworkPolicy(proto.Message): @@ -1186,8 +1475,15 @@ class Provider(proto.Enum): PROVIDER_UNSPECIFIED = 0 CALICO = 1 - provider = proto.Field(proto.ENUM, number=1, enum=Provider,) - enabled = proto.Field(proto.BOOL, number=2,) + provider = proto.Field( + proto.ENUM, + number=1, + enum=Provider, + ) + enabled = proto.Field( + proto.BOOL, + number=2, + ) class IPAllocationPolicy(proto.Message): @@ -1329,20 +1625,62 @@ class IPAllocationPolicy(proto.Message): allocation mode """ - use_ip_aliases = proto.Field(proto.BOOL, number=1,) - create_subnetwork = proto.Field(proto.BOOL, number=2,) - subnetwork_name = proto.Field(proto.STRING, number=3,) - cluster_ipv4_cidr = proto.Field(proto.STRING, number=4,) - node_ipv4_cidr = proto.Field(proto.STRING, number=5,) - services_ipv4_cidr = proto.Field(proto.STRING, number=6,) - cluster_secondary_range_name = proto.Field(proto.STRING, number=7,) - services_secondary_range_name = proto.Field(proto.STRING, number=8,) - cluster_ipv4_cidr_block = proto.Field(proto.STRING, number=9,) - node_ipv4_cidr_block = proto.Field(proto.STRING, number=10,) - services_ipv4_cidr_block = proto.Field(proto.STRING, number=11,) - allow_route_overlap = proto.Field(proto.BOOL, number=12,) - tpu_ipv4_cidr_block = proto.Field(proto.STRING, number=13,) - use_routes = proto.Field(proto.BOOL, number=15,) + use_ip_aliases = proto.Field( + proto.BOOL, + number=1, + ) + create_subnetwork = proto.Field( + proto.BOOL, + number=2, + ) + subnetwork_name = proto.Field( + proto.STRING, + number=3, + ) + cluster_ipv4_cidr = proto.Field( + proto.STRING, + number=4, + ) + node_ipv4_cidr = proto.Field( + proto.STRING, + number=5, + ) + services_ipv4_cidr = proto.Field( + proto.STRING, + number=6, + ) + cluster_secondary_range_name = proto.Field( + proto.STRING, + number=7, + ) + services_secondary_range_name = proto.Field( + proto.STRING, + number=8, + ) + cluster_ipv4_cidr_block = proto.Field( + proto.STRING, + number=9, + ) + node_ipv4_cidr_block = proto.Field( + proto.STRING, + number=10, + ) + services_ipv4_cidr_block = proto.Field( + proto.STRING, + number=11, + ) + allow_route_overlap = proto.Field( + proto.BOOL, + number=12, + ) + tpu_ipv4_cidr_block = proto.Field( + proto.STRING, + number=13, + ) + use_routes = proto.Field( + proto.BOOL, + number=15, + ) class BinaryAuthorization(proto.Message): @@ -1355,7 +1693,10 @@ class BinaryAuthorization(proto.Message): validated by Google Binauthz. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class PodSecurityPolicyConfig(proto.Message): @@ -1368,7 +1709,10 @@ class PodSecurityPolicyConfig(proto.Message): under a PodSecurityPolicy to be created. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class AuthenticatorGroupsConfig(proto.Message): @@ -1385,8 +1729,14 @@ class AuthenticatorGroupsConfig(proto.Message): be used. Only relevant if enabled = true. """ - enabled = proto.Field(proto.BOOL, number=1,) - security_group = proto.Field(proto.STRING, number=2,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) + security_group = proto.Field( + proto.STRING, + number=2, + ) class ClusterTelemetry(proto.Message): @@ -1404,7 +1754,11 @@ class Type(proto.Enum): ENABLED = 2 SYSTEM_ONLY = 3 - type_ = proto.Field(proto.ENUM, number=1, enum=Type,) + type_ = proto.Field( + proto.ENUM, + number=1, + enum=Type, + ) class Cluster(proto.Message): @@ -1529,7 +1883,7 @@ class Cluster(proto.Message): master/node upgrades are disabled. Alpha enabled clusters are automatically deleted thirty days after creation. - resource_labels (Sequence[google.cloud.container_v1beta1.types.Cluster.ResourceLabelsEntry]): + resource_labels (Mapping[str, str]): The resource labels for the cluster to use to annotate any related Google Compute Engine resources. @@ -1732,114 +2086,311 @@ class Status(proto.Enum): ERROR = 5 DEGRADED = 6 - name = proto.Field(proto.STRING, number=1,) - description = proto.Field(proto.STRING, number=2,) - initial_node_count = proto.Field(proto.INT32, number=3,) - node_config = proto.Field(proto.MESSAGE, number=4, message="NodeConfig",) - master_auth = proto.Field(proto.MESSAGE, number=5, message="MasterAuth",) - logging_service = proto.Field(proto.STRING, number=6,) - monitoring_service = proto.Field(proto.STRING, number=7,) - network = proto.Field(proto.STRING, number=8,) - cluster_ipv4_cidr = proto.Field(proto.STRING, number=9,) - addons_config = proto.Field(proto.MESSAGE, number=10, message="AddonsConfig",) - subnetwork = proto.Field(proto.STRING, number=11,) - node_pools = proto.RepeatedField(proto.MESSAGE, number=12, message="NodePool",) - locations = proto.RepeatedField(proto.STRING, number=13,) - enable_kubernetes_alpha = proto.Field(proto.BOOL, number=14,) - resource_labels = proto.MapField(proto.STRING, proto.STRING, number=15,) - label_fingerprint = proto.Field(proto.STRING, number=16,) - legacy_abac = proto.Field(proto.MESSAGE, number=18, message="LegacyAbac",) - network_policy = proto.Field(proto.MESSAGE, number=19, message="NetworkPolicy",) + name = proto.Field( + proto.STRING, + number=1, + ) + description = proto.Field( + proto.STRING, + number=2, + ) + initial_node_count = proto.Field( + proto.INT32, + number=3, + ) + node_config = proto.Field( + proto.MESSAGE, + number=4, + message="NodeConfig", + ) + master_auth = proto.Field( + proto.MESSAGE, + number=5, + message="MasterAuth", + ) + logging_service = proto.Field( + proto.STRING, + number=6, + ) + monitoring_service = proto.Field( + proto.STRING, + number=7, + ) + network = proto.Field( + proto.STRING, + number=8, + ) + cluster_ipv4_cidr = proto.Field( + proto.STRING, + number=9, + ) + addons_config = proto.Field( + proto.MESSAGE, + number=10, + message="AddonsConfig", + ) + subnetwork = proto.Field( + proto.STRING, + number=11, + ) + node_pools = proto.RepeatedField( + proto.MESSAGE, + number=12, + message="NodePool", + ) + locations = proto.RepeatedField( + proto.STRING, + number=13, + ) + enable_kubernetes_alpha = proto.Field( + proto.BOOL, + number=14, + ) + resource_labels = proto.MapField( + proto.STRING, + proto.STRING, + number=15, + ) + label_fingerprint = proto.Field( + proto.STRING, + number=16, + ) + legacy_abac = proto.Field( + proto.MESSAGE, + number=18, + message="LegacyAbac", + ) + network_policy = proto.Field( + proto.MESSAGE, + number=19, + message="NetworkPolicy", + ) ip_allocation_policy = proto.Field( - proto.MESSAGE, number=20, message="IPAllocationPolicy", + proto.MESSAGE, + number=20, + message="IPAllocationPolicy", ) master_authorized_networks_config = proto.Field( - proto.MESSAGE, number=22, message="MasterAuthorizedNetworksConfig", + proto.MESSAGE, + number=22, + message="MasterAuthorizedNetworksConfig", ) maintenance_policy = proto.Field( - proto.MESSAGE, number=23, message="MaintenancePolicy", + proto.MESSAGE, + number=23, + message="MaintenancePolicy", ) binary_authorization = proto.Field( - proto.MESSAGE, number=24, message="BinaryAuthorization", + proto.MESSAGE, + number=24, + message="BinaryAuthorization", ) pod_security_policy_config = proto.Field( - proto.MESSAGE, number=25, message="PodSecurityPolicyConfig", + proto.MESSAGE, + number=25, + message="PodSecurityPolicyConfig", + ) + autoscaling = proto.Field( + proto.MESSAGE, + number=26, + message="ClusterAutoscaling", + ) + network_config = proto.Field( + proto.MESSAGE, + number=27, + message="NetworkConfig", + ) + private_cluster = proto.Field( + proto.BOOL, + number=28, + ) + master_ipv4_cidr_block = proto.Field( + proto.STRING, + number=29, ) - autoscaling = proto.Field(proto.MESSAGE, number=26, message="ClusterAutoscaling",) - network_config = proto.Field(proto.MESSAGE, number=27, message="NetworkConfig",) - private_cluster = proto.Field(proto.BOOL, number=28,) - master_ipv4_cidr_block = proto.Field(proto.STRING, number=29,) default_max_pods_constraint = proto.Field( - proto.MESSAGE, number=30, message="MaxPodsConstraint", + proto.MESSAGE, + number=30, + message="MaxPodsConstraint", ) resource_usage_export_config = proto.Field( - proto.MESSAGE, number=33, message="ResourceUsageExportConfig", + proto.MESSAGE, + number=33, + message="ResourceUsageExportConfig", ) authenticator_groups_config = proto.Field( - proto.MESSAGE, number=34, message="AuthenticatorGroupsConfig", + proto.MESSAGE, + number=34, + message="AuthenticatorGroupsConfig", ) private_cluster_config = proto.Field( - proto.MESSAGE, number=37, message="PrivateClusterConfig", + proto.MESSAGE, + number=37, + message="PrivateClusterConfig", ) vertical_pod_autoscaling = proto.Field( - proto.MESSAGE, number=39, message="VerticalPodAutoscaling", + proto.MESSAGE, + number=39, + message="VerticalPodAutoscaling", + ) + shielded_nodes = proto.Field( + proto.MESSAGE, + number=40, + message="ShieldedNodes", + ) + release_channel = proto.Field( + proto.MESSAGE, + number=41, + message="ReleaseChannel", ) - shielded_nodes = proto.Field(proto.MESSAGE, number=40, message="ShieldedNodes",) - release_channel = proto.Field(proto.MESSAGE, number=41, message="ReleaseChannel",) workload_identity_config = proto.Field( - proto.MESSAGE, number=43, message="WorkloadIdentityConfig", + proto.MESSAGE, + number=43, + message="WorkloadIdentityConfig", ) workload_certificates = proto.Field( - proto.MESSAGE, number=52, message="WorkloadCertificates", + proto.MESSAGE, + number=52, + message="WorkloadCertificates", ) mesh_certificates = proto.Field( - proto.MESSAGE, number=67, message="MeshCertificates", + proto.MESSAGE, + number=67, + message="MeshCertificates", ) cluster_telemetry = proto.Field( - proto.MESSAGE, number=46, message="ClusterTelemetry", + proto.MESSAGE, + number=46, + message="ClusterTelemetry", + ) + tpu_config = proto.Field( + proto.MESSAGE, + number=47, + message="TpuConfig", ) - tpu_config = proto.Field(proto.MESSAGE, number=47, message="TpuConfig",) notification_config = proto.Field( - proto.MESSAGE, number=49, message="NotificationConfig", + proto.MESSAGE, + number=49, + message="NotificationConfig", ) confidential_nodes = proto.Field( - proto.MESSAGE, number=50, message="ConfidentialNodes", + proto.MESSAGE, + number=50, + message="ConfidentialNodes", ) identity_service_config = proto.Field( - proto.MESSAGE, number=54, message="IdentityServiceConfig", - ) - self_link = proto.Field(proto.STRING, number=100,) - zone = proto.Field(proto.STRING, number=101,) - endpoint = proto.Field(proto.STRING, number=102,) - initial_cluster_version = proto.Field(proto.STRING, number=103,) - current_master_version = proto.Field(proto.STRING, number=104,) - current_node_version = proto.Field(proto.STRING, number=105,) - create_time = proto.Field(proto.STRING, number=106,) - status = proto.Field(proto.ENUM, number=107, enum=Status,) - status_message = proto.Field(proto.STRING, number=108,) - node_ipv4_cidr_size = proto.Field(proto.INT32, number=109,) - services_ipv4_cidr = proto.Field(proto.STRING, number=110,) - instance_group_urls = proto.RepeatedField(proto.STRING, number=111,) - current_node_count = proto.Field(proto.INT32, number=112,) - expire_time = proto.Field(proto.STRING, number=113,) - location = proto.Field(proto.STRING, number=114,) - enable_tpu = proto.Field(proto.BOOL, number=115,) - tpu_ipv4_cidr_block = proto.Field(proto.STRING, number=116,) + proto.MESSAGE, + number=54, + message="IdentityServiceConfig", + ) + self_link = proto.Field( + proto.STRING, + number=100, + ) + zone = proto.Field( + proto.STRING, + number=101, + ) + endpoint = proto.Field( + proto.STRING, + number=102, + ) + initial_cluster_version = proto.Field( + proto.STRING, + number=103, + ) + current_master_version = proto.Field( + proto.STRING, + number=104, + ) + current_node_version = proto.Field( + proto.STRING, + number=105, + ) + create_time = proto.Field( + proto.STRING, + number=106, + ) + status = proto.Field( + proto.ENUM, + number=107, + enum=Status, + ) + status_message = proto.Field( + proto.STRING, + number=108, + ) + node_ipv4_cidr_size = proto.Field( + proto.INT32, + number=109, + ) + services_ipv4_cidr = proto.Field( + proto.STRING, + number=110, + ) + instance_group_urls = proto.RepeatedField( + proto.STRING, + number=111, + ) + current_node_count = proto.Field( + proto.INT32, + number=112, + ) + expire_time = proto.Field( + proto.STRING, + number=113, + ) + location = proto.Field( + proto.STRING, + number=114, + ) + enable_tpu = proto.Field( + proto.BOOL, + number=115, + ) + tpu_ipv4_cidr_block = proto.Field( + proto.STRING, + number=116, + ) database_encryption = proto.Field( - proto.MESSAGE, number=38, message="DatabaseEncryption", + proto.MESSAGE, + number=38, + message="DatabaseEncryption", ) conditions = proto.RepeatedField( - proto.MESSAGE, number=118, message="StatusCondition", + proto.MESSAGE, + number=118, + message="StatusCondition", + ) + master = proto.Field( + proto.MESSAGE, + number=124, + message="Master", + ) + autopilot = proto.Field( + proto.MESSAGE, + number=128, + message="Autopilot", + ) + id = proto.Field( + proto.STRING, + number=129, ) - master = proto.Field(proto.MESSAGE, number=124, message="Master",) - autopilot = proto.Field(proto.MESSAGE, number=128, message="Autopilot",) - id = proto.Field(proto.STRING, number=129,) node_pool_defaults = proto.Field( - proto.MESSAGE, number=131, optional=True, message="NodePoolDefaults", + proto.MESSAGE, + number=131, + optional=True, + message="NodePoolDefaults", + ) + logging_config = proto.Field( + proto.MESSAGE, + number=132, + message="LoggingConfig", ) - logging_config = proto.Field(proto.MESSAGE, number=132, message="LoggingConfig",) monitoring_config = proto.Field( - proto.MESSAGE, number=133, message="MonitoringConfig", + proto.MESSAGE, + number=133, + message="MonitoringConfig", ) @@ -1853,7 +2404,9 @@ class NodePoolDefaults(proto.Message): """ node_config_defaults = proto.Field( - proto.MESSAGE, number=1, message="NodeConfigDefaults", + proto.MESSAGE, + number=1, + message="NodeConfigDefaults", ) @@ -1866,7 +2419,11 @@ class NodeConfigDefaults(proto.Message): Riptide) options. """ - gcfs_config = proto.Field(proto.MESSAGE, number=1, message="GcfsConfig",) + gcfs_config = proto.Field( + proto.MESSAGE, + number=1, + message="GcfsConfig", + ) class ClusterUpdate(proto.Message): @@ -2035,97 +2592,188 @@ class ClusterUpdate(proto.Message): configuration. """ - desired_node_version = proto.Field(proto.STRING, number=4,) - desired_monitoring_service = proto.Field(proto.STRING, number=5,) + desired_node_version = proto.Field( + proto.STRING, + number=4, + ) + desired_monitoring_service = proto.Field( + proto.STRING, + number=5, + ) desired_addons_config = proto.Field( - proto.MESSAGE, number=6, message="AddonsConfig", + proto.MESSAGE, + number=6, + message="AddonsConfig", + ) + desired_node_pool_id = proto.Field( + proto.STRING, + number=7, + ) + desired_image_type = proto.Field( + proto.STRING, + number=8, ) - desired_node_pool_id = proto.Field(proto.STRING, number=7,) - desired_image_type = proto.Field(proto.STRING, number=8,) desired_node_pool_autoscaling = proto.Field( - proto.MESSAGE, number=9, message="NodePoolAutoscaling", + proto.MESSAGE, + number=9, + message="NodePoolAutoscaling", + ) + desired_locations = proto.RepeatedField( + proto.STRING, + number=10, ) - desired_locations = proto.RepeatedField(proto.STRING, number=10,) desired_master_authorized_networks_config = proto.Field( - proto.MESSAGE, number=12, message="MasterAuthorizedNetworksConfig", + proto.MESSAGE, + number=12, + message="MasterAuthorizedNetworksConfig", ) desired_pod_security_policy_config = proto.Field( - proto.MESSAGE, number=14, message="PodSecurityPolicyConfig", + proto.MESSAGE, + number=14, + message="PodSecurityPolicyConfig", ) desired_cluster_autoscaling = proto.Field( - proto.MESSAGE, number=15, message="ClusterAutoscaling", + proto.MESSAGE, + number=15, + message="ClusterAutoscaling", ) desired_binary_authorization = proto.Field( - proto.MESSAGE, number=16, message="BinaryAuthorization", + proto.MESSAGE, + number=16, + message="BinaryAuthorization", + ) + desired_logging_service = proto.Field( + proto.STRING, + number=19, ) - desired_logging_service = proto.Field(proto.STRING, number=19,) desired_resource_usage_export_config = proto.Field( - proto.MESSAGE, number=21, message="ResourceUsageExportConfig", + proto.MESSAGE, + number=21, + message="ResourceUsageExportConfig", ) desired_vertical_pod_autoscaling = proto.Field( - proto.MESSAGE, number=22, message="VerticalPodAutoscaling", + proto.MESSAGE, + number=22, + message="VerticalPodAutoscaling", ) desired_private_cluster_config = proto.Field( - proto.MESSAGE, number=25, message="PrivateClusterConfig", + proto.MESSAGE, + number=25, + message="PrivateClusterConfig", ) desired_intra_node_visibility_config = proto.Field( - proto.MESSAGE, number=26, message="IntraNodeVisibilityConfig", + proto.MESSAGE, + number=26, + message="IntraNodeVisibilityConfig", ) desired_default_snat_status = proto.Field( - proto.MESSAGE, number=28, message="DefaultSnatStatus", + proto.MESSAGE, + number=28, + message="DefaultSnatStatus", ) desired_cluster_telemetry = proto.Field( - proto.MESSAGE, number=30, message="ClusterTelemetry", + proto.MESSAGE, + number=30, + message="ClusterTelemetry", ) desired_release_channel = proto.Field( - proto.MESSAGE, number=31, message="ReleaseChannel", + proto.MESSAGE, + number=31, + message="ReleaseChannel", + ) + desired_tpu_config = proto.Field( + proto.MESSAGE, + number=38, + message="TpuConfig", ) - desired_tpu_config = proto.Field(proto.MESSAGE, number=38, message="TpuConfig",) desired_l4ilb_subsetting_config = proto.Field( - proto.MESSAGE, number=39, message="ILBSubsettingConfig", + proto.MESSAGE, + number=39, + message="ILBSubsettingConfig", ) desired_datapath_provider = proto.Field( - proto.ENUM, number=50, enum="DatapathProvider", + proto.ENUM, + number=50, + enum="DatapathProvider", ) desired_private_ipv6_google_access = proto.Field( - proto.ENUM, number=51, enum="PrivateIPv6GoogleAccess", + proto.ENUM, + number=51, + enum="PrivateIPv6GoogleAccess", ) desired_notification_config = proto.Field( - proto.MESSAGE, number=55, message="NotificationConfig", + proto.MESSAGE, + number=55, + message="NotificationConfig", + ) + desired_master_version = proto.Field( + proto.STRING, + number=100, + ) + desired_gcfs_config = proto.Field( + proto.MESSAGE, + number=109, + message="GcfsConfig", ) - desired_master_version = proto.Field(proto.STRING, number=100,) - desired_gcfs_config = proto.Field(proto.MESSAGE, number=109, message="GcfsConfig",) desired_database_encryption = proto.Field( - proto.MESSAGE, number=46, message="DatabaseEncryption", + proto.MESSAGE, + number=46, + message="DatabaseEncryption", ) desired_workload_identity_config = proto.Field( - proto.MESSAGE, number=47, message="WorkloadIdentityConfig", + proto.MESSAGE, + number=47, + message="WorkloadIdentityConfig", ) desired_workload_certificates = proto.Field( - proto.MESSAGE, number=61, message="WorkloadCertificates", + proto.MESSAGE, + number=61, + message="WorkloadCertificates", ) desired_mesh_certificates = proto.Field( - proto.MESSAGE, number=67, message="MeshCertificates", + proto.MESSAGE, + number=67, + message="MeshCertificates", ) desired_shielded_nodes = proto.Field( - proto.MESSAGE, number=48, message="ShieldedNodes", + proto.MESSAGE, + number=48, + message="ShieldedNodes", + ) + desired_master = proto.Field( + proto.MESSAGE, + number=52, + message="Master", + ) + desired_dns_config = proto.Field( + proto.MESSAGE, + number=53, + message="DNSConfig", ) - desired_master = proto.Field(proto.MESSAGE, number=52, message="Master",) - desired_dns_config = proto.Field(proto.MESSAGE, number=53, message="DNSConfig",) desired_service_external_ips_config = proto.Field( - proto.MESSAGE, number=60, message="ServiceExternalIPsConfig", + proto.MESSAGE, + number=60, + message="ServiceExternalIPsConfig", ) desired_authenticator_groups_config = proto.Field( - proto.MESSAGE, number=63, message="AuthenticatorGroupsConfig", + proto.MESSAGE, + number=63, + message="AuthenticatorGroupsConfig", ) desired_logging_config = proto.Field( - proto.MESSAGE, number=64, message="LoggingConfig", + proto.MESSAGE, + number=64, + message="LoggingConfig", ) desired_monitoring_config = proto.Field( - proto.MESSAGE, number=65, message="MonitoringConfig", + proto.MESSAGE, + number=65, + message="MonitoringConfig", ) desired_identity_service_config = proto.Field( - proto.MESSAGE, number=66, message="IdentityServiceConfig", + proto.MESSAGE, + number=66, + message="IdentityServiceConfig", ) @@ -2213,25 +2861,72 @@ class Type(proto.Enum): SET_NETWORK_POLICY = 15 SET_MAINTENANCE_POLICY = 16 - name = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - operation_type = proto.Field(proto.ENUM, number=3, enum=Type,) - status = proto.Field(proto.ENUM, number=4, enum=Status,) - detail = proto.Field(proto.STRING, number=8,) - status_message = proto.Field(proto.STRING, number=5,) - self_link = proto.Field(proto.STRING, number=6,) - target_link = proto.Field(proto.STRING, number=7,) - location = proto.Field(proto.STRING, number=9,) - start_time = proto.Field(proto.STRING, number=10,) - end_time = proto.Field(proto.STRING, number=11,) - progress = proto.Field(proto.MESSAGE, number=12, message="OperationProgress",) + name = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + operation_type = proto.Field( + proto.ENUM, + number=3, + enum=Type, + ) + status = proto.Field( + proto.ENUM, + number=4, + enum=Status, + ) + detail = proto.Field( + proto.STRING, + number=8, + ) + status_message = proto.Field( + proto.STRING, + number=5, + ) + self_link = proto.Field( + proto.STRING, + number=6, + ) + target_link = proto.Field( + proto.STRING, + number=7, + ) + location = proto.Field( + proto.STRING, + number=9, + ) + start_time = proto.Field( + proto.STRING, + number=10, + ) + end_time = proto.Field( + proto.STRING, + number=11, + ) + progress = proto.Field( + proto.MESSAGE, + number=12, + message="OperationProgress", + ) cluster_conditions = proto.RepeatedField( - proto.MESSAGE, number=13, message="StatusCondition", + proto.MESSAGE, + number=13, + message="StatusCondition", ) nodepool_conditions = proto.RepeatedField( - proto.MESSAGE, number=14, message="StatusCondition", + proto.MESSAGE, + number=14, + message="StatusCondition", + ) + error = proto.Field( + proto.MESSAGE, + number=15, + message=status_pb2.Status, ) - error = proto.Field(proto.MESSAGE, number=15, message=status_pb2.Status,) class OperationProgress(proto.Message): @@ -2283,15 +2978,45 @@ class Metric(proto.Message): This field is a member of `oneof`_ ``value``. """ - name = proto.Field(proto.STRING, number=1,) - int_value = proto.Field(proto.INT64, number=2, oneof="value",) - double_value = proto.Field(proto.DOUBLE, number=3, oneof="value",) - string_value = proto.Field(proto.STRING, number=4, oneof="value",) + name = proto.Field( + proto.STRING, + number=1, + ) + int_value = proto.Field( + proto.INT64, + number=2, + oneof="value", + ) + double_value = proto.Field( + proto.DOUBLE, + number=3, + oneof="value", + ) + string_value = proto.Field( + proto.STRING, + number=4, + oneof="value", + ) - name = proto.Field(proto.STRING, number=1,) - status = proto.Field(proto.ENUM, number=2, enum="Operation.Status",) - metrics = proto.RepeatedField(proto.MESSAGE, number=3, message=Metric,) - stages = proto.RepeatedField(proto.MESSAGE, number=4, message="OperationProgress",) + name = proto.Field( + proto.STRING, + number=1, + ) + status = proto.Field( + proto.ENUM, + number=2, + enum="Operation.Status", + ) + metrics = proto.RepeatedField( + proto.MESSAGE, + number=3, + message=Metric, + ) + stages = proto.RepeatedField( + proto.MESSAGE, + number=4, + message="OperationProgress", + ) class CreateClusterRequest(proto.Message): @@ -2317,10 +3042,23 @@ class CreateClusterRequest(proto.Message): created. Specified in the format ``projects/*/locations/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster = proto.Field(proto.MESSAGE, number=3, message="Cluster",) - parent = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster = proto.Field( + proto.MESSAGE, + number=3, + message="Cluster", + ) + parent = proto.Field( + proto.STRING, + number=5, + ) class GetClusterRequest(proto.Message): @@ -2348,10 +3086,22 @@ class GetClusterRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=5, + ) class UpdateClusterRequest(proto.Message): @@ -2381,11 +3131,27 @@ class UpdateClusterRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - update = proto.Field(proto.MESSAGE, number=4, message="ClusterUpdate",) - name = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + update = proto.Field( + proto.MESSAGE, + number=4, + message="ClusterUpdate", + ) + name = proto.Field( + proto.STRING, + number=5, + ) class UpdateNodePoolRequest(proto.Message): @@ -2471,29 +3237,83 @@ class UpdateNodePoolRequest(proto.Message): Enable or disable gvnic on the node pool. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - node_version = proto.Field(proto.STRING, number=5,) - image_type = proto.Field(proto.STRING, number=6,) - locations = proto.RepeatedField(proto.STRING, number=13,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + node_version = proto.Field( + proto.STRING, + number=5, + ) + image_type = proto.Field( + proto.STRING, + number=6, + ) + locations = proto.RepeatedField( + proto.STRING, + number=13, + ) workload_metadata_config = proto.Field( - proto.MESSAGE, number=14, message="WorkloadMetadataConfig", + proto.MESSAGE, + number=14, + message="WorkloadMetadataConfig", + ) + name = proto.Field( + proto.STRING, + number=8, ) - name = proto.Field(proto.STRING, number=8,) upgrade_settings = proto.Field( - proto.MESSAGE, number=15, message="NodePool.UpgradeSettings", + proto.MESSAGE, + number=15, + message="NodePool.UpgradeSettings", + ) + tags = proto.Field( + proto.MESSAGE, + number=16, + message="NetworkTags", + ) + taints = proto.Field( + proto.MESSAGE, + number=17, + message="NodeTaints", + ) + labels = proto.Field( + proto.MESSAGE, + number=18, + message="NodeLabels", ) - tags = proto.Field(proto.MESSAGE, number=16, message="NetworkTags",) - taints = proto.Field(proto.MESSAGE, number=17, message="NodeTaints",) - labels = proto.Field(proto.MESSAGE, number=18, message="NodeLabels",) linux_node_config = proto.Field( - proto.MESSAGE, number=19, message="LinuxNodeConfig", + proto.MESSAGE, + number=19, + message="LinuxNodeConfig", + ) + kubelet_config = proto.Field( + proto.MESSAGE, + number=20, + message="NodeKubeletConfig", + ) + gcfs_config = proto.Field( + proto.MESSAGE, + number=22, + message="GcfsConfig", + ) + gvnic = proto.Field( + proto.MESSAGE, + number=29, + message="VirtualNIC", ) - kubelet_config = proto.Field(proto.MESSAGE, number=20, message="NodeKubeletConfig",) - gcfs_config = proto.Field(proto.MESSAGE, number=22, message="GcfsConfig",) - gvnic = proto.Field(proto.MESSAGE, number=29, message="VirtualNIC",) class SetNodePoolAutoscalingRequest(proto.Message): @@ -2529,12 +3349,31 @@ class SetNodePoolAutoscalingRequest(proto.Message): ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - autoscaling = proto.Field(proto.MESSAGE, number=5, message="NodePoolAutoscaling",) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + autoscaling = proto.Field( + proto.MESSAGE, + number=5, + message="NodePoolAutoscaling", + ) + name = proto.Field( + proto.STRING, + number=6, + ) class SetLoggingServiceRequest(proto.Message): @@ -2577,11 +3416,26 @@ class SetLoggingServiceRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - logging_service = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + logging_service = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=5, + ) class SetMonitoringServiceRequest(proto.Message): @@ -2625,11 +3479,26 @@ class SetMonitoringServiceRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - monitoring_service = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + monitoring_service = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class SetAddonsConfigRequest(proto.Message): @@ -2660,11 +3529,27 @@ class SetAddonsConfigRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - addons_config = proto.Field(proto.MESSAGE, number=4, message="AddonsConfig",) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + addons_config = proto.Field( + proto.MESSAGE, + number=4, + message="AddonsConfig", + ) + name = proto.Field( + proto.STRING, + number=6, + ) class SetLocationsRequest(proto.Message): @@ -2701,11 +3586,26 @@ class SetLocationsRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - locations = proto.RepeatedField(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + locations = proto.RepeatedField( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class UpdateMasterRequest(proto.Message): @@ -2746,11 +3646,26 @@ class UpdateMasterRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - master_version = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + master_version = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=7, + ) class SetMasterAuthRequest(proto.Message): @@ -2790,12 +3705,32 @@ class Action(proto.Enum): GENERATE_PASSWORD = 2 SET_USERNAME = 3 - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - action = proto.Field(proto.ENUM, number=4, enum=Action,) - update = proto.Field(proto.MESSAGE, number=5, message="MasterAuth",) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + action = proto.Field( + proto.ENUM, + number=4, + enum=Action, + ) + update = proto.Field( + proto.MESSAGE, + number=5, + message="MasterAuth", + ) + name = proto.Field( + proto.STRING, + number=7, + ) class DeleteClusterRequest(proto.Message): @@ -2823,10 +3758,22 @@ class DeleteClusterRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=4,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=4, + ) class ListClustersRequest(proto.Message): @@ -2850,9 +3797,18 @@ class ListClustersRequest(proto.Message): Location "-" matches all zones and all regions. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - parent = proto.Field(proto.STRING, number=4,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + parent = proto.Field( + proto.STRING, + number=4, + ) class ListClustersResponse(proto.Message): @@ -2867,8 +3823,15 @@ class ListClustersResponse(proto.Message): clusters returned may be missing those zones. """ - clusters = proto.RepeatedField(proto.MESSAGE, number=1, message="Cluster",) - missing_zones = proto.RepeatedField(proto.STRING, number=2,) + clusters = proto.RepeatedField( + proto.MESSAGE, + number=1, + message="Cluster", + ) + missing_zones = proto.RepeatedField( + proto.STRING, + number=2, + ) class GetOperationRequest(proto.Message): @@ -2896,10 +3859,22 @@ class GetOperationRequest(proto.Message): ``projects/*/locations/*/operations/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - operation_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + operation_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=5, + ) class ListOperationsRequest(proto.Message): @@ -2924,9 +3899,18 @@ class ListOperationsRequest(proto.Message): and all regions. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - parent = proto.Field(proto.STRING, number=4,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + parent = proto.Field( + proto.STRING, + number=4, + ) class CancelOperationRequest(proto.Message): @@ -2954,10 +3938,22 @@ class CancelOperationRequest(proto.Message): ``projects/*/locations/*/operations/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - operation_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=4,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + operation_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=4, + ) class ListOperationsResponse(proto.Message): @@ -2974,8 +3970,15 @@ class ListOperationsResponse(proto.Message): operations from those zones. """ - operations = proto.RepeatedField(proto.MESSAGE, number=1, message="Operation",) - missing_zones = proto.RepeatedField(proto.STRING, number=2,) + operations = proto.RepeatedField( + proto.MESSAGE, + number=1, + message="Operation", + ) + missing_zones = proto.RepeatedField( + proto.STRING, + number=2, + ) class GetServerConfigRequest(proto.Message): @@ -2998,9 +4001,18 @@ class GetServerConfigRequest(proto.Message): specified in the format ``projects/*/locations/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - name = proto.Field(proto.STRING, number=4,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + name = proto.Field( + proto.STRING, + number=4, + ) class ServerConfig(proto.Message): @@ -3022,7 +4034,7 @@ class ServerConfig(proto.Message): order. channels (Sequence[google.cloud.container_v1beta1.types.ServerConfig.ReleaseChannelConfig]): List of release channel configurations. - windows_version_maps (Sequence[google.cloud.container_v1beta1.types.ServerConfig.WindowsVersionMapsEntry]): + windows_version_maps (Mapping[str, google.cloud.container_v1beta1.types.WindowsVersions]): Maps of Kubernetes version and supported Windows server versions. """ @@ -3055,28 +4067,64 @@ class AvailableVersion(proto.Message): Reason for availability. """ - version = proto.Field(proto.STRING, number=1,) - reason = proto.Field(proto.STRING, number=2,) - - channel = proto.Field(proto.ENUM, number=1, enum="ReleaseChannel.Channel",) - default_version = proto.Field(proto.STRING, number=2,) + version = proto.Field( + proto.STRING, + number=1, + ) + reason = proto.Field( + proto.STRING, + number=2, + ) + + channel = proto.Field( + proto.ENUM, + number=1, + enum="ReleaseChannel.Channel", + ) + default_version = proto.Field( + proto.STRING, + number=2, + ) available_versions = proto.RepeatedField( proto.MESSAGE, number=3, message="ServerConfig.ReleaseChannelConfig.AvailableVersion", ) - valid_versions = proto.RepeatedField(proto.STRING, number=4,) + valid_versions = proto.RepeatedField( + proto.STRING, + number=4, + ) - default_cluster_version = proto.Field(proto.STRING, number=1,) - valid_node_versions = proto.RepeatedField(proto.STRING, number=3,) - default_image_type = proto.Field(proto.STRING, number=4,) - valid_image_types = proto.RepeatedField(proto.STRING, number=5,) - valid_master_versions = proto.RepeatedField(proto.STRING, number=6,) + default_cluster_version = proto.Field( + proto.STRING, + number=1, + ) + valid_node_versions = proto.RepeatedField( + proto.STRING, + number=3, + ) + default_image_type = proto.Field( + proto.STRING, + number=4, + ) + valid_image_types = proto.RepeatedField( + proto.STRING, + number=5, + ) + valid_master_versions = proto.RepeatedField( + proto.STRING, + number=6, + ) channels = proto.RepeatedField( - proto.MESSAGE, number=9, message=ReleaseChannelConfig, + proto.MESSAGE, + number=9, + message=ReleaseChannelConfig, ) windows_version_maps = proto.MapField( - proto.STRING, proto.MESSAGE, number=10, message="WindowsVersions", + proto.STRING, + proto.MESSAGE, + number=10, + message="WindowsVersions", ) @@ -3100,12 +4148,24 @@ class WindowsVersion(proto.Message): Mainstream support end date """ - image_type = proto.Field(proto.STRING, number=1,) - os_version = proto.Field(proto.STRING, number=2,) - support_end_date = proto.Field(proto.MESSAGE, number=3, message=date_pb2.Date,) + image_type = proto.Field( + proto.STRING, + number=1, + ) + os_version = proto.Field( + proto.STRING, + number=2, + ) + support_end_date = proto.Field( + proto.MESSAGE, + number=3, + message=date_pb2.Date, + ) windows_versions = proto.RepeatedField( - proto.MESSAGE, number=1, message=WindowsVersion, + proto.MESSAGE, + number=1, + message=WindowsVersion, ) @@ -3136,11 +4196,27 @@ class CreateNodePoolRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool = proto.Field(proto.MESSAGE, number=4, message="NodePool",) - parent = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool = proto.Field( + proto.MESSAGE, + number=4, + message="NodePool", + ) + parent = proto.Field( + proto.STRING, + number=6, + ) class DeleteNodePoolRequest(proto.Message): @@ -3172,11 +4248,26 @@ class DeleteNodePoolRequest(proto.Message): ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class ListNodePoolsRequest(proto.Message): @@ -3204,10 +4295,22 @@ class ListNodePoolsRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - parent = proto.Field(proto.STRING, number=5,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + parent = proto.Field( + proto.STRING, + number=5, + ) class GetNodePoolRequest(proto.Message): @@ -3239,11 +4342,26 @@ class GetNodePoolRequest(proto.Message): ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class NodePool(proto.Message): @@ -3368,8 +4486,14 @@ class UpgradeSettings(proto.Message): status is Ready. """ - max_surge = proto.Field(proto.INT32, number=1,) - max_unavailable = proto.Field(proto.INT32, number=2,) + max_surge = proto.Field( + proto.INT32, + number=1, + ) + max_unavailable = proto.Field( + proto.INT32, + number=2, + ) class PlacementPolicy(proto.Message): r"""PlacementPolicy defines the placement policy used by the node @@ -3385,29 +4509,89 @@ class Type(proto.Enum): TYPE_UNSPECIFIED = 0 COMPACT = 1 - type_ = proto.Field(proto.ENUM, number=1, enum="NodePool.PlacementPolicy.Type",) - - name = proto.Field(proto.STRING, number=1,) - config = proto.Field(proto.MESSAGE, number=2, message="NodeConfig",) - initial_node_count = proto.Field(proto.INT32, number=3,) - locations = proto.RepeatedField(proto.STRING, number=13,) - network_config = proto.Field(proto.MESSAGE, number=14, message="NodeNetworkConfig",) - self_link = proto.Field(proto.STRING, number=100,) - version = proto.Field(proto.STRING, number=101,) - instance_group_urls = proto.RepeatedField(proto.STRING, number=102,) - status = proto.Field(proto.ENUM, number=103, enum=Status,) - status_message = proto.Field(proto.STRING, number=104,) - autoscaling = proto.Field(proto.MESSAGE, number=4, message="NodePoolAutoscaling",) - management = proto.Field(proto.MESSAGE, number=5, message="NodeManagement",) + type_ = proto.Field( + proto.ENUM, + number=1, + enum="NodePool.PlacementPolicy.Type", + ) + + name = proto.Field( + proto.STRING, + number=1, + ) + config = proto.Field( + proto.MESSAGE, + number=2, + message="NodeConfig", + ) + initial_node_count = proto.Field( + proto.INT32, + number=3, + ) + locations = proto.RepeatedField( + proto.STRING, + number=13, + ) + network_config = proto.Field( + proto.MESSAGE, + number=14, + message="NodeNetworkConfig", + ) + self_link = proto.Field( + proto.STRING, + number=100, + ) + version = proto.Field( + proto.STRING, + number=101, + ) + instance_group_urls = proto.RepeatedField( + proto.STRING, + number=102, + ) + status = proto.Field( + proto.ENUM, + number=103, + enum=Status, + ) + status_message = proto.Field( + proto.STRING, + number=104, + ) + autoscaling = proto.Field( + proto.MESSAGE, + number=4, + message="NodePoolAutoscaling", + ) + management = proto.Field( + proto.MESSAGE, + number=5, + message="NodeManagement", + ) max_pods_constraint = proto.Field( - proto.MESSAGE, number=6, message="MaxPodsConstraint", + proto.MESSAGE, + number=6, + message="MaxPodsConstraint", ) conditions = proto.RepeatedField( - proto.MESSAGE, number=105, message="StatusCondition", + proto.MESSAGE, + number=105, + message="StatusCondition", + ) + pod_ipv4_cidr_size = proto.Field( + proto.INT32, + number=7, + ) + upgrade_settings = proto.Field( + proto.MESSAGE, + number=107, + message=UpgradeSettings, + ) + placement_policy = proto.Field( + proto.MESSAGE, + number=108, + message=PlacementPolicy, ) - pod_ipv4_cidr_size = proto.Field(proto.INT32, number=7,) - upgrade_settings = proto.Field(proto.MESSAGE, number=107, message=UpgradeSettings,) - placement_policy = proto.Field(proto.MESSAGE, number=108, message=PlacementPolicy,) class NodeManagement(proto.Message): @@ -3426,10 +4610,18 @@ class NodeManagement(proto.Message): pool. """ - auto_upgrade = proto.Field(proto.BOOL, number=1,) - auto_repair = proto.Field(proto.BOOL, number=2,) + auto_upgrade = proto.Field( + proto.BOOL, + number=1, + ) + auto_repair = proto.Field( + proto.BOOL, + number=2, + ) upgrade_options = proto.Field( - proto.MESSAGE, number=10, message="AutoUpgradeOptions", + proto.MESSAGE, + number=10, + message="AutoUpgradeOptions", ) @@ -3448,8 +4640,14 @@ class AutoUpgradeOptions(proto.Message): commence with the description of the upgrade. """ - auto_upgrade_start_time = proto.Field(proto.STRING, number=1,) - description = proto.Field(proto.STRING, number=2,) + auto_upgrade_start_time = proto.Field( + proto.STRING, + number=1, + ) + description = proto.Field( + proto.STRING, + number=2, + ) class MaintenancePolicy(proto.Message): @@ -3470,8 +4668,15 @@ class MaintenancePolicy(proto.Message): policy. """ - window = proto.Field(proto.MESSAGE, number=1, message="MaintenanceWindow",) - resource_version = proto.Field(proto.STRING, number=3,) + window = proto.Field( + proto.MESSAGE, + number=1, + message="MaintenanceWindow", + ) + resource_version = proto.Field( + proto.STRING, + number=3, + ) class MaintenanceWindow(proto.Message): @@ -3499,20 +4704,29 @@ class MaintenanceWindow(proto.Message): occur at any time. This field is a member of `oneof`_ ``policy``. - maintenance_exclusions (Sequence[google.cloud.container_v1beta1.types.MaintenanceWindow.MaintenanceExclusionsEntry]): + maintenance_exclusions (Mapping[str, google.cloud.container_v1beta1.types.TimeWindow]): Exceptions to maintenance window. Non-emergency maintenance should not occur in these windows. """ daily_maintenance_window = proto.Field( - proto.MESSAGE, number=2, oneof="policy", message="DailyMaintenanceWindow", + proto.MESSAGE, + number=2, + oneof="policy", + message="DailyMaintenanceWindow", ) recurring_window = proto.Field( - proto.MESSAGE, number=3, oneof="policy", message="RecurringTimeWindow", + proto.MESSAGE, + number=3, + oneof="policy", + message="RecurringTimeWindow", ) maintenance_exclusions = proto.MapField( - proto.STRING, proto.MESSAGE, number=4, message="TimeWindow", + proto.STRING, + proto.MESSAGE, + number=4, + message="TimeWindow", ) @@ -3535,10 +4749,21 @@ class TimeWindow(proto.Message): """ maintenance_exclusion_options = proto.Field( - proto.MESSAGE, number=3, oneof="options", message="MaintenanceExclusionOptions", + proto.MESSAGE, + number=3, + oneof="options", + message="MaintenanceExclusionOptions", + ) + start_time = proto.Field( + proto.MESSAGE, + number=1, + message=timestamp_pb2.Timestamp, + ) + end_time = proto.Field( + proto.MESSAGE, + number=2, + message=timestamp_pb2.Timestamp, ) - start_time = proto.Field(proto.MESSAGE, number=1, message=timestamp_pb2.Timestamp,) - end_time = proto.Field(proto.MESSAGE, number=2, message=timestamp_pb2.Timestamp,) class MaintenanceExclusionOptions(proto.Message): @@ -3556,7 +4781,11 @@ class Scope(proto.Enum): NO_MINOR_UPGRADES = 1 NO_MINOR_OR_NODE_UPGRADES = 2 - scope = proto.Field(proto.ENUM, number=1, enum=Scope,) + scope = proto.Field( + proto.ENUM, + number=1, + enum=Scope, + ) class RecurringTimeWindow(proto.Message): @@ -3606,8 +4835,15 @@ class RecurringTimeWindow(proto.Message): SECONDLY are not supported. """ - window = proto.Field(proto.MESSAGE, number=1, message="TimeWindow",) - recurrence = proto.Field(proto.STRING, number=2,) + window = proto.Field( + proto.MESSAGE, + number=1, + message="TimeWindow", + ) + recurrence = proto.Field( + proto.STRING, + number=2, + ) class DailyMaintenanceWindow(proto.Message): @@ -3623,8 +4859,14 @@ class DailyMaintenanceWindow(proto.Message): chosen to be smallest possible in the given scenario. """ - start_time = proto.Field(proto.STRING, number=2,) - duration = proto.Field(proto.STRING, number=3,) + start_time = proto.Field( + proto.STRING, + number=2, + ) + duration = proto.Field( + proto.STRING, + number=3, + ) class SetNodePoolManagementRequest(proto.Message): @@ -3660,12 +4902,31 @@ class SetNodePoolManagementRequest(proto.Message): format ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - management = proto.Field(proto.MESSAGE, number=5, message="NodeManagement",) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + management = proto.Field( + proto.MESSAGE, + number=5, + message="NodeManagement", + ) + name = proto.Field( + proto.STRING, + number=7, + ) class SetNodePoolSizeRequest(proto.Message): @@ -3700,12 +4961,30 @@ class SetNodePoolSizeRequest(proto.Message): ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - node_count = proto.Field(proto.INT32, number=5,) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + node_count = proto.Field( + proto.INT32, + number=5, + ) + name = proto.Field( + proto.STRING, + number=7, + ) class RollbackNodePoolUpgradeRequest(proto.Message): @@ -3739,11 +5018,26 @@ class RollbackNodePoolUpgradeRequest(proto.Message): ``projects/*/locations/*/clusters/*/nodePools/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - node_pool_id = proto.Field(proto.STRING, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + node_pool_id = proto.Field( + proto.STRING, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class ListNodePoolsResponse(proto.Message): @@ -3754,7 +5048,11 @@ class ListNodePoolsResponse(proto.Message): A list of node pools for a cluster. """ - node_pools = proto.RepeatedField(proto.MESSAGE, number=1, message="NodePool",) + node_pools = proto.RepeatedField( + proto.MESSAGE, + number=1, + message="NodePool", + ) class ClusterAutoscaling(proto.Message): @@ -3787,15 +5085,29 @@ class AutoscalingProfile(proto.Enum): OPTIMIZE_UTILIZATION = 1 BALANCED = 2 - enable_node_autoprovisioning = proto.Field(proto.BOOL, number=1,) + enable_node_autoprovisioning = proto.Field( + proto.BOOL, + number=1, + ) resource_limits = proto.RepeatedField( - proto.MESSAGE, number=2, message="ResourceLimit", + proto.MESSAGE, + number=2, + message="ResourceLimit", + ) + autoscaling_profile = proto.Field( + proto.ENUM, + number=3, + enum=AutoscalingProfile, ) - autoscaling_profile = proto.Field(proto.ENUM, number=3, enum=AutoscalingProfile,) autoprovisioning_node_pool_defaults = proto.Field( - proto.MESSAGE, number=4, message="AutoprovisioningNodePoolDefaults", + proto.MESSAGE, + number=4, + message="AutoprovisioningNodePoolDefaults", + ) + autoprovisioning_locations = proto.RepeatedField( + proto.STRING, + number=5, ) - autoprovisioning_locations = proto.RepeatedField(proto.STRING, number=5,) class AutoprovisioningNodePoolDefaults(proto.Message): @@ -3867,20 +5179,49 @@ class AutoprovisioningNodePoolDefaults(proto.Message): The image type to use for NAP created node. """ - oauth_scopes = proto.RepeatedField(proto.STRING, number=1,) - service_account = proto.Field(proto.STRING, number=2,) + oauth_scopes = proto.RepeatedField( + proto.STRING, + number=1, + ) + service_account = proto.Field( + proto.STRING, + number=2, + ) upgrade_settings = proto.Field( - proto.MESSAGE, number=3, message="NodePool.UpgradeSettings", + proto.MESSAGE, + number=3, + message="NodePool.UpgradeSettings", + ) + management = proto.Field( + proto.MESSAGE, + number=4, + message="NodeManagement", + ) + min_cpu_platform = proto.Field( + proto.STRING, + number=5, + ) + disk_size_gb = proto.Field( + proto.INT32, + number=6, + ) + disk_type = proto.Field( + proto.STRING, + number=7, ) - management = proto.Field(proto.MESSAGE, number=4, message="NodeManagement",) - min_cpu_platform = proto.Field(proto.STRING, number=5,) - disk_size_gb = proto.Field(proto.INT32, number=6,) - disk_type = proto.Field(proto.STRING, number=7,) shielded_instance_config = proto.Field( - proto.MESSAGE, number=8, message="ShieldedInstanceConfig", + proto.MESSAGE, + number=8, + message="ShieldedInstanceConfig", + ) + boot_disk_kms_key = proto.Field( + proto.STRING, + number=9, + ) + image_type = proto.Field( + proto.STRING, + number=10, ) - boot_disk_kms_key = proto.Field(proto.STRING, number=9,) - image_type = proto.Field(proto.STRING, number=10,) class ResourceLimit(proto.Message): @@ -3899,9 +5240,18 @@ class ResourceLimit(proto.Message): cluster. """ - resource_type = proto.Field(proto.STRING, number=1,) - minimum = proto.Field(proto.INT64, number=2,) - maximum = proto.Field(proto.INT64, number=3,) + resource_type = proto.Field( + proto.STRING, + number=1, + ) + minimum = proto.Field( + proto.INT64, + number=2, + ) + maximum = proto.Field( + proto.INT64, + number=3, + ) class NodePoolAutoscaling(proto.Message): @@ -3923,10 +5273,22 @@ class NodePoolAutoscaling(proto.Message): Can this node pool be deleted automatically. """ - enabled = proto.Field(proto.BOOL, number=1,) - min_node_count = proto.Field(proto.INT32, number=2,) - max_node_count = proto.Field(proto.INT32, number=3,) - autoprovisioned = proto.Field(proto.BOOL, number=4,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) + min_node_count = proto.Field( + proto.INT32, + number=2, + ) + max_node_count = proto.Field( + proto.INT32, + number=3, + ) + autoprovisioned = proto.Field( + proto.BOOL, + number=4, + ) class SetLabelsRequest(proto.Message): @@ -3950,7 +5312,7 @@ class SetLabelsRequest(proto.Message): Required. Deprecated. The name of the cluster. This field has been deprecated and replaced by the name field. - resource_labels (Sequence[google.cloud.container_v1beta1.types.SetLabelsRequest.ResourceLabelsEntry]): + resource_labels (Mapping[str, str]): Required. The labels to set for that cluster. label_fingerprint (str): Required. The fingerprint of the previous set of labels for @@ -3966,12 +5328,31 @@ class SetLabelsRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - resource_labels = proto.MapField(proto.STRING, proto.STRING, number=4,) - label_fingerprint = proto.Field(proto.STRING, number=5,) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + resource_labels = proto.MapField( + proto.STRING, + proto.STRING, + number=4, + ) + label_fingerprint = proto.Field( + proto.STRING, + number=5, + ) + name = proto.Field( + proto.STRING, + number=7, + ) class SetLegacyAbacRequest(proto.Message): @@ -4003,11 +5384,26 @@ class SetLegacyAbacRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - enabled = proto.Field(proto.BOOL, number=4,) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + enabled = proto.Field( + proto.BOOL, + number=4, + ) + name = proto.Field( + proto.STRING, + number=6, + ) class StartIPRotationRequest(proto.Message): @@ -4040,11 +5436,26 @@ class StartIPRotationRequest(proto.Message): rotation. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=6,) - rotate_credentials = proto.Field(proto.BOOL, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=6, + ) + rotate_credentials = proto.Field( + proto.BOOL, + number=7, + ) class CompleteIPRotationRequest(proto.Message): @@ -4073,10 +5484,22 @@ class CompleteIPRotationRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - name = proto.Field(proto.STRING, number=7,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + name = proto.Field( + proto.STRING, + number=7, + ) class AcceleratorConfig(proto.Message): @@ -4096,9 +5519,18 @@ class AcceleratorConfig(proto.Message): guide `__. """ - accelerator_count = proto.Field(proto.INT64, number=1,) - accelerator_type = proto.Field(proto.STRING, number=2,) - gpu_partition_size = proto.Field(proto.STRING, number=3,) + accelerator_count = proto.Field( + proto.INT64, + number=1, + ) + accelerator_type = proto.Field( + proto.STRING, + number=2, + ) + gpu_partition_size = proto.Field( + proto.STRING, + number=3, + ) class ManagedPrometheusConfig(proto.Message): @@ -4110,7 +5542,10 @@ class ManagedPrometheusConfig(proto.Message): Enable Managed Collection. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class WorkloadMetadataConfig(proto.Message): @@ -4144,8 +5579,16 @@ class Mode(proto.Enum): GCE_METADATA = 1 GKE_METADATA = 2 - node_metadata = proto.Field(proto.ENUM, number=1, enum=NodeMetadata,) - mode = proto.Field(proto.ENUM, number=2, enum=Mode,) + node_metadata = proto.Field( + proto.ENUM, + number=1, + enum=NodeMetadata, + ) + mode = proto.Field( + proto.ENUM, + number=2, + enum=Mode, + ) class SetNetworkPolicyRequest(proto.Message): @@ -4177,11 +5620,27 @@ class SetNetworkPolicyRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) - network_policy = proto.Field(proto.MESSAGE, number=4, message="NetworkPolicy",) - name = proto.Field(proto.STRING, number=6,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) + network_policy = proto.Field( + proto.MESSAGE, + number=4, + message="NetworkPolicy", + ) + name = proto.Field( + proto.STRING, + number=6, + ) class SetMaintenancePolicyRequest(proto.Message): @@ -4209,13 +5668,27 @@ class SetMaintenancePolicyRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - project_id = proto.Field(proto.STRING, number=1,) - zone = proto.Field(proto.STRING, number=2,) - cluster_id = proto.Field(proto.STRING, number=3,) + project_id = proto.Field( + proto.STRING, + number=1, + ) + zone = proto.Field( + proto.STRING, + number=2, + ) + cluster_id = proto.Field( + proto.STRING, + number=3, + ) maintenance_policy = proto.Field( - proto.MESSAGE, number=4, message="MaintenancePolicy", + proto.MESSAGE, + number=4, + message="MaintenancePolicy", + ) + name = proto.Field( + proto.STRING, + number=5, ) - name = proto.Field(proto.STRING, number=5,) class ListLocationsRequest(proto.Message): @@ -4228,7 +5701,10 @@ class ListLocationsRequest(proto.Message): Specified in the format ``projects/*``. """ - parent = proto.Field(proto.STRING, number=1,) + parent = proto.Field( + proto.STRING, + number=1, + ) class ListLocationsResponse(proto.Message): @@ -4250,8 +5726,15 @@ class ListLocationsResponse(proto.Message): def raw_page(self): return self - locations = proto.RepeatedField(proto.MESSAGE, number=1, message="Location",) - next_page_token = proto.Field(proto.STRING, number=2,) + locations = proto.RepeatedField( + proto.MESSAGE, + number=1, + message="Location", + ) + next_page_token = proto.Field( + proto.STRING, + number=2, + ) class Location(proto.Message): @@ -4276,9 +5759,19 @@ class LocationType(proto.Enum): ZONE = 1 REGION = 2 - type_ = proto.Field(proto.ENUM, number=1, enum=LocationType,) - name = proto.Field(proto.STRING, number=2,) - recommended = proto.Field(proto.BOOL, number=3,) + type_ = proto.Field( + proto.ENUM, + number=1, + enum=LocationType, + ) + name = proto.Field( + proto.STRING, + number=2, + ) + recommended = proto.Field( + proto.BOOL, + number=3, + ) class StatusCondition(proto.Message): @@ -4307,9 +5800,20 @@ class Code(proto.Enum): CLOUD_KMS_KEY_ERROR = 7 CA_EXPIRING = 9 - code = proto.Field(proto.ENUM, number=1, enum=Code,) - message = proto.Field(proto.STRING, number=2,) - canonical_code = proto.Field(proto.ENUM, number=3, enum=code_pb2.Code,) + code = proto.Field( + proto.ENUM, + number=1, + enum=Code, + ) + message = proto.Field( + proto.STRING, + number=2, + ) + canonical_code = proto.Field( + proto.ENUM, + number=3, + enum=code_pb2.Code, + ) class NetworkConfig(proto.Message): @@ -4358,20 +5862,46 @@ class NetworkConfig(proto.Message): not. """ - network = proto.Field(proto.STRING, number=1,) - subnetwork = proto.Field(proto.STRING, number=2,) - enable_intra_node_visibility = proto.Field(proto.BOOL, number=5,) + network = proto.Field( + proto.STRING, + number=1, + ) + subnetwork = proto.Field( + proto.STRING, + number=2, + ) + enable_intra_node_visibility = proto.Field( + proto.BOOL, + number=5, + ) default_snat_status = proto.Field( - proto.MESSAGE, number=7, message="DefaultSnatStatus", + proto.MESSAGE, + number=7, + message="DefaultSnatStatus", + ) + enable_l4ilb_subsetting = proto.Field( + proto.BOOL, + number=10, + ) + datapath_provider = proto.Field( + proto.ENUM, + number=11, + enum="DatapathProvider", ) - enable_l4ilb_subsetting = proto.Field(proto.BOOL, number=10,) - datapath_provider = proto.Field(proto.ENUM, number=11, enum="DatapathProvider",) private_ipv6_google_access = proto.Field( - proto.ENUM, number=12, enum="PrivateIPv6GoogleAccess", + proto.ENUM, + number=12, + enum="PrivateIPv6GoogleAccess", + ) + dns_config = proto.Field( + proto.MESSAGE, + number=13, + message="DNSConfig", ) - dns_config = proto.Field(proto.MESSAGE, number=13, message="DNSConfig",) service_external_ips_config = proto.Field( - proto.MESSAGE, number=15, message="ServiceExternalIPsConfig", + proto.MESSAGE, + number=15, + message="ServiceExternalIPsConfig", ) @@ -4384,7 +5914,10 @@ class ServiceExternalIPsConfig(proto.Message): allowed or not. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class ListUsableSubnetworksRequest(proto.Message): @@ -4414,10 +5947,22 @@ class ListUsableSubnetworksRequest(proto.Message): requests to get the next page of results. """ - parent = proto.Field(proto.STRING, number=1,) - filter = proto.Field(proto.STRING, number=2,) - page_size = proto.Field(proto.INT32, number=3,) - page_token = proto.Field(proto.STRING, number=4,) + parent = proto.Field( + proto.STRING, + number=1, + ) + filter = proto.Field( + proto.STRING, + number=2, + ) + page_size = proto.Field( + proto.INT32, + number=3, + ) + page_token = proto.Field( + proto.STRING, + number=4, + ) class ListUsableSubnetworksResponse(proto.Message): @@ -4441,9 +5986,14 @@ def raw_page(self): return self subnetworks = proto.RepeatedField( - proto.MESSAGE, number=1, message="UsableSubnetwork", + proto.MESSAGE, + number=1, + message="UsableSubnetwork", + ) + next_page_token = proto.Field( + proto.STRING, + number=2, ) - next_page_token = proto.Field(proto.STRING, number=2,) class UsableSubnetworkSecondaryRange(proto.Message): @@ -4470,9 +6020,19 @@ class Status(proto.Enum): IN_USE_SHAREABLE_POD = 3 IN_USE_MANAGED_POD = 4 - range_name = proto.Field(proto.STRING, number=1,) - ip_cidr_range = proto.Field(proto.STRING, number=2,) - status = proto.Field(proto.ENUM, number=3, enum=Status,) + range_name = proto.Field( + proto.STRING, + number=1, + ) + ip_cidr_range = proto.Field( + proto.STRING, + number=2, + ) + status = proto.Field( + proto.ENUM, + number=3, + enum=Status, + ) class UsableSubnetwork(proto.Message): @@ -4501,13 +6061,27 @@ class UsableSubnetwork(proto.Message): message will be given by status_message. """ - subnetwork = proto.Field(proto.STRING, number=1,) - network = proto.Field(proto.STRING, number=2,) - ip_cidr_range = proto.Field(proto.STRING, number=3,) + subnetwork = proto.Field( + proto.STRING, + number=1, + ) + network = proto.Field( + proto.STRING, + number=2, + ) + ip_cidr_range = proto.Field( + proto.STRING, + number=3, + ) secondary_ip_ranges = proto.RepeatedField( - proto.MESSAGE, number=4, message="UsableSubnetworkSecondaryRange", + proto.MESSAGE, + number=4, + message="UsableSubnetworkSecondaryRange", + ) + status_message = proto.Field( + proto.STRING, + number=5, ) - status_message = proto.Field(proto.STRING, number=5,) class VerticalPodAutoscaling(proto.Message): @@ -4520,7 +6094,10 @@ class VerticalPodAutoscaling(proto.Message): Enables vertical pod autoscaling. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class DefaultSnatStatus(proto.Message): @@ -4532,7 +6109,10 @@ class DefaultSnatStatus(proto.Message): Disables cluster default sNAT rules. """ - disabled = proto.Field(proto.BOOL, number=1,) + disabled = proto.Field( + proto.BOOL, + number=1, + ) class IntraNodeVisibilityConfig(proto.Message): @@ -4545,7 +6125,10 @@ class IntraNodeVisibilityConfig(proto.Message): cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class ILBSubsettingConfig(proto.Message): @@ -4557,7 +6140,10 @@ class ILBSubsettingConfig(proto.Message): Enables l4 ILB subsetting for this cluster """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class DNSConfig(proto.Message): @@ -4590,9 +6176,20 @@ class DNSScope(proto.Enum): CLUSTER_SCOPE = 1 VPC_SCOPE = 2 - cluster_dns = proto.Field(proto.ENUM, number=1, enum=Provider,) - cluster_dns_scope = proto.Field(proto.ENUM, number=2, enum=DNSScope,) - cluster_dns_domain = proto.Field(proto.STRING, number=3,) + cluster_dns = proto.Field( + proto.ENUM, + number=1, + enum=Provider, + ) + cluster_dns_scope = proto.Field( + proto.ENUM, + number=2, + enum=DNSScope, + ) + cluster_dns_domain = proto.Field( + proto.STRING, + number=3, + ) class MaxPodsConstraint(proto.Message): @@ -4604,7 +6201,10 @@ class MaxPodsConstraint(proto.Message): per node. """ - max_pods_per_node = proto.Field(proto.INT64, number=1,) + max_pods_per_node = proto.Field( + proto.INT64, + number=1, + ) class WorkloadIdentityConfig(proto.Message): @@ -4623,9 +6223,18 @@ class WorkloadIdentityConfig(proto.Message): provider. """ - identity_namespace = proto.Field(proto.STRING, number=1,) - workload_pool = proto.Field(proto.STRING, number=2,) - identity_provider = proto.Field(proto.STRING, number=3,) + identity_namespace = proto.Field( + proto.STRING, + number=1, + ) + workload_pool = proto.Field( + proto.STRING, + number=2, + ) + identity_provider = proto.Field( + proto.STRING, + number=3, + ) class WorkloadCertificates(proto.Message): @@ -4648,7 +6257,9 @@ class WorkloadCertificates(proto.Message): """ enable_certificates = proto.Field( - proto.MESSAGE, number=1, message=wrappers_pb2.BoolValue, + proto.MESSAGE, + number=1, + message=wrappers_pb2.BoolValue, ) @@ -4672,7 +6283,9 @@ class MeshCertificates(proto.Message): """ enable_certificates = proto.Field( - proto.MESSAGE, number=1, message=wrappers_pb2.BoolValue, + proto.MESSAGE, + number=1, + message=wrappers_pb2.BoolValue, ) @@ -4694,8 +6307,15 @@ class State(proto.Enum): ENCRYPTED = 1 DECRYPTED = 2 - state = proto.Field(proto.ENUM, number=2, enum=State,) - key_name = proto.Field(proto.STRING, number=1,) + state = proto.Field( + proto.ENUM, + number=2, + enum=State, + ) + key_name = proto.Field( + proto.STRING, + number=1, + ) class ResourceUsageExportConfig(proto.Message): @@ -4724,7 +6344,10 @@ class BigQueryDestination(proto.Message): The ID of a BigQuery Dataset. """ - dataset_id = proto.Field(proto.STRING, number=1,) + dataset_id = proto.Field( + proto.STRING, + number=1, + ) class ConsumptionMeteringConfig(proto.Message): r"""Parameters for controlling consumption metering. @@ -4737,14 +6360,24 @@ class ConsumptionMeteringConfig(proto.Message): consumption records. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) bigquery_destination = proto.Field( - proto.MESSAGE, number=1, message=BigQueryDestination, + proto.MESSAGE, + number=1, + message=BigQueryDestination, + ) + enable_network_egress_metering = proto.Field( + proto.BOOL, + number=2, ) - enable_network_egress_metering = proto.Field(proto.BOOL, number=2,) consumption_metering_config = proto.Field( - proto.MESSAGE, number=3, message=ConsumptionMeteringConfig, + proto.MESSAGE, + number=3, + message=ConsumptionMeteringConfig, ) @@ -4757,7 +6390,10 @@ class ShieldedNodes(proto.Message): on all nodes in this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class VirtualNIC(proto.Message): @@ -4769,7 +6405,10 @@ class VirtualNIC(proto.Message): node pool. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class GetOpenIDConfigRequest(proto.Message): @@ -4784,7 +6423,10 @@ class GetOpenIDConfigRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - parent = proto.Field(proto.STRING, number=1,) + parent = proto.Field( + proto.STRING, + number=1, + ) class GetOpenIDConfigResponse(proto.Message): @@ -4809,13 +6451,34 @@ class GetOpenIDConfigResponse(proto.Message): Supported grant types. """ - issuer = proto.Field(proto.STRING, number=1,) - jwks_uri = proto.Field(proto.STRING, number=2,) - response_types_supported = proto.RepeatedField(proto.STRING, number=3,) - subject_types_supported = proto.RepeatedField(proto.STRING, number=4,) - id_token_signing_alg_values_supported = proto.RepeatedField(proto.STRING, number=5,) - claims_supported = proto.RepeatedField(proto.STRING, number=6,) - grant_types = proto.RepeatedField(proto.STRING, number=7,) + issuer = proto.Field( + proto.STRING, + number=1, + ) + jwks_uri = proto.Field( + proto.STRING, + number=2, + ) + response_types_supported = proto.RepeatedField( + proto.STRING, + number=3, + ) + subject_types_supported = proto.RepeatedField( + proto.STRING, + number=4, + ) + id_token_signing_alg_values_supported = proto.RepeatedField( + proto.STRING, + number=5, + ) + claims_supported = proto.RepeatedField( + proto.STRING, + number=6, + ) + grant_types = proto.RepeatedField( + proto.STRING, + number=7, + ) class GetJSONWebKeysRequest(proto.Message): @@ -4831,7 +6494,10 @@ class GetJSONWebKeysRequest(proto.Message): ``projects/*/locations/*/clusters/*``. """ - parent = proto.Field(proto.STRING, number=1,) + parent = proto.Field( + proto.STRING, + number=1, + ) class Jwk(proto.Message): @@ -4858,15 +6524,42 @@ class Jwk(proto.Message): Used for ECDSA keys. """ - kty = proto.Field(proto.STRING, number=1,) - alg = proto.Field(proto.STRING, number=2,) - use = proto.Field(proto.STRING, number=3,) - kid = proto.Field(proto.STRING, number=4,) - n = proto.Field(proto.STRING, number=5,) - e = proto.Field(proto.STRING, number=6,) - x = proto.Field(proto.STRING, number=7,) - y = proto.Field(proto.STRING, number=8,) - crv = proto.Field(proto.STRING, number=9,) + kty = proto.Field( + proto.STRING, + number=1, + ) + alg = proto.Field( + proto.STRING, + number=2, + ) + use = proto.Field( + proto.STRING, + number=3, + ) + kid = proto.Field( + proto.STRING, + number=4, + ) + n = proto.Field( + proto.STRING, + number=5, + ) + e = proto.Field( + proto.STRING, + number=6, + ) + x = proto.Field( + proto.STRING, + number=7, + ) + y = proto.Field( + proto.STRING, + number=8, + ) + crv = proto.Field( + proto.STRING, + number=9, + ) class GetJSONWebKeysResponse(proto.Message): @@ -4879,7 +6572,11 @@ class GetJSONWebKeysResponse(proto.Message): cluster to sign token requests. """ - keys = proto.RepeatedField(proto.MESSAGE, number=1, message="Jwk",) + keys = proto.RepeatedField( + proto.MESSAGE, + number=1, + message="Jwk", + ) class ReleaseChannel(proto.Message): @@ -4902,7 +6599,11 @@ class Channel(proto.Enum): REGULAR = 2 STABLE = 3 - channel = proto.Field(proto.ENUM, number=1, enum=Channel,) + channel = proto.Field( + proto.ENUM, + number=1, + enum=Channel, + ) class TpuConfig(proto.Message): @@ -4920,14 +6621,22 @@ class TpuConfig(proto.Message): VPC. """ - enabled = proto.Field(proto.BOOL, number=1,) - use_service_networking = proto.Field(proto.BOOL, number=2,) - ipv4_cidr_block = proto.Field(proto.STRING, number=3,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) + use_service_networking = proto.Field( + proto.BOOL, + number=2, + ) + ipv4_cidr_block = proto.Field( + proto.STRING, + number=3, + ) class Master(proto.Message): - r"""Master is the configuration for components on master. - """ + r"""Master is the configuration for components on master.""" class Autopilot(proto.Message): @@ -4939,7 +6648,10 @@ class Autopilot(proto.Message): Enable Autopilot """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class NotificationConfig(proto.Message): @@ -4976,10 +6688,18 @@ class PubSub(proto.Message): event types will be sent """ - enabled = proto.Field(proto.BOOL, number=1,) - topic = proto.Field(proto.STRING, number=2,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) + topic = proto.Field( + proto.STRING, + number=2, + ) filter = proto.Field( - proto.MESSAGE, number=3, message="NotificationConfig.Filter", + proto.MESSAGE, + number=3, + message="NotificationConfig.Filter", ) class Filter(proto.Message): @@ -4995,10 +6715,16 @@ class Filter(proto.Message): """ event_type = proto.RepeatedField( - proto.ENUM, number=1, enum="NotificationConfig.EventType", + proto.ENUM, + number=1, + enum="NotificationConfig.EventType", ) - pubsub = proto.Field(proto.MESSAGE, number=1, message=PubSub,) + pubsub = proto.Field( + proto.MESSAGE, + number=1, + message=PubSub, + ) class ConfidentialNodes(proto.Message): @@ -5011,7 +6737,10 @@ class ConfidentialNodes(proto.Message): for all nodes in this cluster. """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class UpgradeEvent(proto.Message): @@ -5035,14 +6764,32 @@ class UpgradeEvent(proto.Message): of the node pool. """ - resource_type = proto.Field(proto.ENUM, number=1, enum="UpgradeResourceType",) - operation = proto.Field(proto.STRING, number=2,) + resource_type = proto.Field( + proto.ENUM, + number=1, + enum="UpgradeResourceType", + ) + operation = proto.Field( + proto.STRING, + number=2, + ) operation_start_time = proto.Field( - proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, + proto.MESSAGE, + number=3, + message=timestamp_pb2.Timestamp, + ) + current_version = proto.Field( + proto.STRING, + number=4, + ) + target_version = proto.Field( + proto.STRING, + number=5, + ) + resource = proto.Field( + proto.STRING, + number=6, ) - current_version = proto.Field(proto.STRING, number=4,) - target_version = proto.Field(proto.STRING, number=5,) - resource = proto.Field(proto.STRING, number=6,) class UpgradeAvailableEvent(proto.Message): @@ -5064,11 +6811,29 @@ class UpgradeAvailableEvent(proto.Message): Windows node versions info. """ - version = proto.Field(proto.STRING, number=1,) - resource_type = proto.Field(proto.ENUM, number=2, enum="UpgradeResourceType",) - release_channel = proto.Field(proto.MESSAGE, number=3, message="ReleaseChannel",) - resource = proto.Field(proto.STRING, number=4,) - windows_versions = proto.Field(proto.MESSAGE, number=5, message="WindowsVersions",) + version = proto.Field( + proto.STRING, + number=1, + ) + resource_type = proto.Field( + proto.ENUM, + number=2, + enum="UpgradeResourceType", + ) + release_channel = proto.Field( + proto.MESSAGE, + number=3, + message="ReleaseChannel", + ) + resource = proto.Field( + proto.STRING, + number=4, + ) + windows_versions = proto.Field( + proto.MESSAGE, + number=5, + message="WindowsVersions", + ) class SecurityBulletinEvent(proto.Message): @@ -5116,16 +6881,46 @@ class SecurityBulletinEvent(proto.Message): their clusters safe. """ - resource_type_affected = proto.Field(proto.STRING, number=1,) - bulletin_id = proto.Field(proto.STRING, number=2,) - cve_ids = proto.RepeatedField(proto.STRING, number=3,) - severity = proto.Field(proto.STRING, number=4,) - bulletin_uri = proto.Field(proto.STRING, number=5,) - brief_description = proto.Field(proto.STRING, number=6,) - affected_supported_minors = proto.RepeatedField(proto.STRING, number=7,) - patched_versions = proto.RepeatedField(proto.STRING, number=8,) - suggested_upgrade_target = proto.Field(proto.STRING, number=9,) - manual_steps_required = proto.Field(proto.BOOL, number=10,) + resource_type_affected = proto.Field( + proto.STRING, + number=1, + ) + bulletin_id = proto.Field( + proto.STRING, + number=2, + ) + cve_ids = proto.RepeatedField( + proto.STRING, + number=3, + ) + severity = proto.Field( + proto.STRING, + number=4, + ) + bulletin_uri = proto.Field( + proto.STRING, + number=5, + ) + brief_description = proto.Field( + proto.STRING, + number=6, + ) + affected_supported_minors = proto.RepeatedField( + proto.STRING, + number=7, + ) + patched_versions = proto.RepeatedField( + proto.STRING, + number=8, + ) + suggested_upgrade_target = proto.Field( + proto.STRING, + number=9, + ) + manual_steps_required = proto.Field( + proto.BOOL, + number=10, + ) class IdentityServiceConfig(proto.Message): @@ -5139,7 +6934,10 @@ class IdentityServiceConfig(proto.Message): component """ - enabled = proto.Field(proto.BOOL, number=1,) + enabled = proto.Field( + proto.BOOL, + number=1, + ) class LoggingConfig(proto.Message): @@ -5151,7 +6949,9 @@ class LoggingConfig(proto.Message): """ component_config = proto.Field( - proto.MESSAGE, number=1, message="LoggingComponentConfig", + proto.MESSAGE, + number=1, + message="LoggingComponentConfig", ) @@ -5171,7 +6971,11 @@ class Component(proto.Enum): SYSTEM_COMPONENTS = 1 WORKLOADS = 2 - enable_components = proto.RepeatedField(proto.ENUM, number=1, enum=Component,) + enable_components = proto.RepeatedField( + proto.ENUM, + number=1, + enum=Component, + ) class MonitoringConfig(proto.Message): @@ -5186,10 +6990,14 @@ class MonitoringConfig(proto.Message): """ component_config = proto.Field( - proto.MESSAGE, number=1, message="MonitoringComponentConfig", + proto.MESSAGE, + number=1, + message="MonitoringComponentConfig", ) managed_prometheus_config = proto.Field( - proto.MESSAGE, number=2, message="ManagedPrometheusConfig", + proto.MESSAGE, + number=2, + message="ManagedPrometheusConfig", ) @@ -5209,7 +7017,11 @@ class Component(proto.Enum): SYSTEM_COMPONENTS = 1 WORKLOADS = 2 - enable_components = proto.RepeatedField(proto.ENUM, number=1, enum=Component,) + enable_components = proto.RepeatedField( + proto.ENUM, + number=1, + enum=Component, + ) __all__ = tuple(sorted(__protobuf__.manifest)) diff --git a/noxfile.py b/noxfile.py index 2a2001c4..7c1742d2 100644 --- a/noxfile.py +++ b/noxfile.py @@ -17,19 +17,45 @@ # Generated by synthtool. DO NOT EDIT! from __future__ import absolute_import + import os import pathlib import shutil +import warnings import nox - -BLACK_VERSION = "black==19.10b0" -BLACK_PATHS = ["docs", "google", "tests", "noxfile.py", "setup.py"] +BLACK_VERSION = "black==22.3.0" +ISORT_VERSION = "isort==5.10.1" +LINT_PATHS = ["docs", "google", "tests", "noxfile.py", "setup.py"] DEFAULT_PYTHON_VERSION = "3.8" -SYSTEM_TEST_PYTHON_VERSIONS = ["3.8"] + UNIT_TEST_PYTHON_VERSIONS = ["3.6", "3.7", "3.8", "3.9", "3.10"] +UNIT_TEST_STANDARD_DEPENDENCIES = [ + "mock", + "asyncmock", + "pytest", + "pytest-cov", + "pytest-asyncio", +] +UNIT_TEST_EXTERNAL_DEPENDENCIES = [] +UNIT_TEST_LOCAL_DEPENDENCIES = [] +UNIT_TEST_DEPENDENCIES = [] +UNIT_TEST_EXTRAS = [] +UNIT_TEST_EXTRAS_BY_PYTHON = {} + +SYSTEM_TEST_PYTHON_VERSIONS = ["3.8"] +SYSTEM_TEST_STANDARD_DEPENDENCIES = [ + "mock", + "pytest", + "google-cloud-testutils", +] +SYSTEM_TEST_EXTERNAL_DEPENDENCIES = [] +SYSTEM_TEST_LOCAL_DEPENDENCIES = [] +SYSTEM_TEST_DEPENDENCIES = [] +SYSTEM_TEST_EXTRAS = [] +SYSTEM_TEST_EXTRAS_BY_PYTHON = {} CURRENT_DIRECTORY = pathlib.Path(__file__).parent.absolute() @@ -57,7 +83,9 @@ def lint(session): """ session.install("flake8", BLACK_VERSION) session.run( - "black", "--check", *BLACK_PATHS, + "black", + "--check", + *LINT_PATHS, ) session.run("flake8", "google", "tests") @@ -67,7 +95,28 @@ def blacken(session): """Run black. Format code to uniform standard.""" session.install(BLACK_VERSION) session.run( - "black", *BLACK_PATHS, + "black", + *LINT_PATHS, + ) + + +@nox.session(python=DEFAULT_PYTHON_VERSION) +def format(session): + """ + Run isort to sort imports. Then run black + to format code to uniform standard. + """ + session.install(BLACK_VERSION, ISORT_VERSION) + # Use the --fss option to sort imports using strict alphabetical order. + # See https://pycqa.github.io/isort/docs/configuration/options.html#force-sort-within-sections + session.run( + "isort", + "--fss", + *LINT_PATHS, + ) + session.run( + "black", + *LINT_PATHS, ) @@ -78,23 +127,41 @@ def lint_setup_py(session): session.run("python", "setup.py", "check", "--restructuredtext", "--strict") +def install_unittest_dependencies(session, *constraints): + standard_deps = UNIT_TEST_STANDARD_DEPENDENCIES + UNIT_TEST_DEPENDENCIES + session.install(*standard_deps, *constraints) + + if UNIT_TEST_EXTERNAL_DEPENDENCIES: + warnings.warn( + "'unit_test_external_dependencies' is deprecated. Instead, please " + "use 'unit_test_dependencies' or 'unit_test_local_dependencies'.", + DeprecationWarning, + ) + session.install(*UNIT_TEST_EXTERNAL_DEPENDENCIES, *constraints) + + if UNIT_TEST_LOCAL_DEPENDENCIES: + session.install(*UNIT_TEST_LOCAL_DEPENDENCIES, *constraints) + + if UNIT_TEST_EXTRAS_BY_PYTHON: + extras = UNIT_TEST_EXTRAS_BY_PYTHON.get(session.python, []) + elif UNIT_TEST_EXTRAS: + extras = UNIT_TEST_EXTRAS + else: + extras = [] + + if extras: + session.install("-e", f".[{','.join(extras)}]", *constraints) + else: + session.install("-e", ".", *constraints) + + def default(session): # Install all test dependencies, then install this package in-place. constraints_path = str( CURRENT_DIRECTORY / "testing" / f"constraints-{session.python}.txt" ) - session.install( - "mock", - "asyncmock", - "pytest", - "pytest-cov", - "pytest-asyncio", - "-c", - constraints_path, - ) - - session.install("-e", ".", "-c", constraints_path) + install_unittest_dependencies(session, "-c", constraints_path) # Run py.test against the unit tests. session.run( @@ -118,6 +185,35 @@ def unit(session): default(session) +def install_systemtest_dependencies(session, *constraints): + + # Use pre-release gRPC for system tests. + session.install("--pre", "grpcio") + + session.install(*SYSTEM_TEST_STANDARD_DEPENDENCIES, *constraints) + + if SYSTEM_TEST_EXTERNAL_DEPENDENCIES: + session.install(*SYSTEM_TEST_EXTERNAL_DEPENDENCIES, *constraints) + + if SYSTEM_TEST_LOCAL_DEPENDENCIES: + session.install("-e", *SYSTEM_TEST_LOCAL_DEPENDENCIES, *constraints) + + if SYSTEM_TEST_DEPENDENCIES: + session.install("-e", *SYSTEM_TEST_DEPENDENCIES, *constraints) + + if SYSTEM_TEST_EXTRAS_BY_PYTHON: + extras = SYSTEM_TEST_EXTRAS_BY_PYTHON.get(session.python, []) + elif SYSTEM_TEST_EXTRAS: + extras = SYSTEM_TEST_EXTRAS + else: + extras = [] + + if extras: + session.install("-e", f".[{','.join(extras)}]", *constraints) + else: + session.install("-e", ".", *constraints) + + @nox.session(python=SYSTEM_TEST_PYTHON_VERSIONS) def system(session): """Run the system test suite.""" @@ -140,13 +236,7 @@ def system(session): if not system_test_exists and not system_test_folder_exists: session.skip("System tests were not found") - # Use pre-release gRPC for system tests. - session.install("--pre", "grpcio") - - # Install all test dependencies, then install this package into the - # virtualenv's dist-packages. - session.install("mock", "pytest", "google-cloud-testutils", "-c", constraints_path) - session.install("-e", ".", "-c", constraints_path) + install_systemtest_dependencies(session, "-c", constraints_path) # Run py.test against the system tests. if system_test_exists: diff --git a/samples/generated_samples/snippet_metadata_container_v1.json b/samples/generated_samples/snippet_metadata_container_v1.json index af9c4a26..e2cd36d9 100644 --- a/samples/generated_samples/snippet_metadata_container_v1.json +++ b/samples/generated_samples/snippet_metadata_container_v1.json @@ -1,16 +1,72 @@ { + "clientLibrary": { + "apis": [ + { + "id": "google.container.v1", + "version": "v1" + } + ], + "language": "PYTHON", + "name": "google-cloud-container" + }, "snippets": [ { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.cancel_operation", "method": { + "fullName": "google.container.v1.ClusterManager.CancelOperation", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CancelOperation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.CancelOperationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "operation_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "shortName": "cancel_operation" }, + "description": "Sample for CancelOperation", "file": "container_v1_generated_cluster_manager_cancel_operation_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_CancelOperation_async", "segments": [ { @@ -41,18 +97,65 @@ "end": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_cancel_operation_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.cancel_operation", "method": { + "fullName": "google.container.v1.ClusterManager.CancelOperation", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CancelOperation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.CancelOperationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "operation_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "shortName": "cancel_operation" }, + "description": "Sample for CancelOperation", "file": "container_v1_generated_cluster_manager_cancel_operation_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_CancelOperation_sync", "segments": [ { @@ -83,19 +186,67 @@ "end": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_cancel_operation_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.complete_ip_rotation", "method": { + "fullName": "google.container.v1.ClusterManager.CompleteIPRotation", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CompleteIPRotation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.CompleteIPRotationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "complete_ip_rotation" }, + "description": "Sample for CompleteIPRotation", "file": "container_v1_generated_cluster_manager_complete_ip_rotation_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_CompleteIPRotation_async", "segments": [ { @@ -128,18 +279,66 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_complete_ip_rotation_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.complete_ip_rotation", "method": { + "fullName": "google.container.v1.ClusterManager.CompleteIPRotation", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CompleteIPRotation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.CompleteIPRotationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "complete_ip_rotation" }, + "description": "Sample for CompleteIPRotation", "file": "container_v1_generated_cluster_manager_complete_ip_rotation_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_CompleteIPRotation_sync", "segments": [ { @@ -172,19 +371,67 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_complete_ip_rotation_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.create_cluster", "method": { + "fullName": "google.container.v1.ClusterManager.CreateCluster", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CreateCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.CreateClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster", + "type": "google.cloud.container_v1.types.Cluster" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "create_cluster" }, + "description": "Sample for CreateCluster", "file": "container_v1_generated_cluster_manager_create_cluster_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_CreateCluster_async", "segments": [ { @@ -217,18 +464,66 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_create_cluster_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.create_cluster", "method": { + "fullName": "google.container.v1.ClusterManager.CreateCluster", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CreateCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.CreateClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster", + "type": "google.cloud.container_v1.types.Cluster" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "create_cluster" }, + "description": "Sample for CreateCluster", "file": "container_v1_generated_cluster_manager_create_cluster_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_CreateCluster_sync", "segments": [ { @@ -261,19 +556,71 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_create_cluster_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.create_node_pool", "method": { + "fullName": "google.container.v1.ClusterManager.CreateNodePool", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CreateNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.CreateNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool", + "type": "google.cloud.container_v1.types.NodePool" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "create_node_pool" }, + "description": "Sample for CreateNodePool", "file": "container_v1_generated_cluster_manager_create_node_pool_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_CreateNodePool_async", "segments": [ { @@ -306,18 +653,70 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_create_node_pool_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.create_node_pool", "method": { + "fullName": "google.container.v1.ClusterManager.CreateNodePool", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CreateNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.CreateNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool", + "type": "google.cloud.container_v1.types.NodePool" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "create_node_pool" }, + "description": "Sample for CreateNodePool", "file": "container_v1_generated_cluster_manager_create_node_pool_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_CreateNodePool_sync", "segments": [ { @@ -350,19 +749,67 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_create_node_pool_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.delete_cluster", "method": { + "fullName": "google.container.v1.ClusterManager.DeleteCluster", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "DeleteCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.DeleteClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "delete_cluster" }, + "description": "Sample for DeleteCluster", "file": "container_v1_generated_cluster_manager_delete_cluster_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_DeleteCluster_async", "segments": [ { @@ -395,18 +842,66 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_delete_cluster_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.delete_cluster", "method": { + "fullName": "google.container.v1.ClusterManager.DeleteCluster", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "DeleteCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.DeleteClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "delete_cluster" }, + "description": "Sample for DeleteCluster", "file": "container_v1_generated_cluster_manager_delete_cluster_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_DeleteCluster_sync", "segments": [ { @@ -439,19 +934,71 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_delete_cluster_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.delete_node_pool", "method": { + "fullName": "google.container.v1.ClusterManager.DeleteNodePool", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "DeleteNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.DeleteNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "delete_node_pool" }, + "description": "Sample for DeleteNodePool", "file": "container_v1_generated_cluster_manager_delete_node_pool_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_DeleteNodePool_async", "segments": [ { @@ -484,18 +1031,70 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_delete_node_pool_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.delete_node_pool", "method": { + "fullName": "google.container.v1.ClusterManager.DeleteNodePool", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "DeleteNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.DeleteNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "delete_node_pool" }, + "description": "Sample for DeleteNodePool", "file": "container_v1_generated_cluster_manager_delete_node_pool_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_DeleteNodePool_sync", "segments": [ { @@ -528,19 +1127,67 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_delete_node_pool_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.get_cluster", "method": { + "fullName": "google.container.v1.ClusterManager.GetCluster", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.GetClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Cluster", + "shortName": "get_cluster" }, + "description": "Sample for GetCluster", "file": "container_v1_generated_cluster_manager_get_cluster_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_GetCluster_async", "segments": [ { @@ -573,18 +1220,66 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_get_cluster_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.get_cluster", "method": { + "fullName": "google.container.v1.ClusterManager.GetCluster", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.GetClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Cluster", + "shortName": "get_cluster" }, + "description": "Sample for GetCluster", "file": "container_v1_generated_cluster_manager_get_cluster_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_GetCluster_sync", "segments": [ { @@ -617,19 +1312,51 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_get_cluster_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.get_json_web_keys", "method": { + "fullName": "google.container.v1.ClusterManager.GetJSONWebKeys", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetJSONWebKeys" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.GetJSONWebKeysRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.GetJSONWebKeysResponse", + "shortName": "get_json_web_keys" }, + "description": "Sample for GetJSONWebKeys", "file": "container_v1_generated_cluster_manager_get_json_web_keys_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_GetJSONWebKeys_async", "segments": [ { @@ -662,18 +1389,50 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_get_json_web_keys_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.get_json_web_keys", "method": { + "fullName": "google.container.v1.ClusterManager.GetJSONWebKeys", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetJSONWebKeys" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.GetJSONWebKeysRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.GetJSONWebKeysResponse", + "shortName": "get_json_web_keys" }, + "description": "Sample for GetJSONWebKeys", "file": "container_v1_generated_cluster_manager_get_json_web_keys_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_GetJSONWebKeys_sync", "segments": [ { @@ -706,19 +1465,71 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_get_json_web_keys_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.get_node_pool", "method": { + "fullName": "google.container.v1.ClusterManager.GetNodePool", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.GetNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.NodePool", + "shortName": "get_node_pool" }, + "description": "Sample for GetNodePool", "file": "container_v1_generated_cluster_manager_get_node_pool_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_GetNodePool_async", "segments": [ { @@ -751,18 +1562,70 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_get_node_pool_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.get_node_pool", "method": { + "fullName": "google.container.v1.ClusterManager.GetNodePool", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.GetNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.NodePool", + "shortName": "get_node_pool" }, + "description": "Sample for GetNodePool", "file": "container_v1_generated_cluster_manager_get_node_pool_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_GetNodePool_sync", "segments": [ { @@ -795,19 +1658,67 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_get_node_pool_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.get_operation", "method": { + "fullName": "google.container.v1.ClusterManager.GetOperation", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetOperation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.GetOperationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "operation_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "get_operation" }, + "description": "Sample for GetOperation", "file": "container_v1_generated_cluster_manager_get_operation_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_GetOperation_async", "segments": [ { @@ -840,18 +1751,66 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_get_operation_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.get_operation", "method": { + "fullName": "google.container.v1.ClusterManager.GetOperation", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetOperation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.GetOperationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "operation_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "get_operation" }, + "description": "Sample for GetOperation", "file": "container_v1_generated_cluster_manager_get_operation_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_GetOperation_sync", "segments": [ { @@ -884,19 +1843,63 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_get_operation_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.get_server_config", "method": { + "fullName": "google.container.v1.ClusterManager.GetServerConfig", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetServerConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.GetServerConfigRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.ServerConfig", + "shortName": "get_server_config" }, + "description": "Sample for GetServerConfig", "file": "container_v1_generated_cluster_manager_get_server_config_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_GetServerConfig_async", "segments": [ { @@ -929,18 +1932,62 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_get_server_config_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.get_server_config", "method": { + "fullName": "google.container.v1.ClusterManager.GetServerConfig", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetServerConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.GetServerConfigRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.ServerConfig", + "shortName": "get_server_config" }, + "description": "Sample for GetServerConfig", "file": "container_v1_generated_cluster_manager_get_server_config_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_GetServerConfig_sync", "segments": [ { @@ -973,19 +2020,63 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_get_server_config_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.list_clusters", "method": { + "fullName": "google.container.v1.ClusterManager.ListClusters", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListClusters" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.ListClustersRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.ListClustersResponse", + "shortName": "list_clusters" }, + "description": "Sample for ListClusters", "file": "container_v1_generated_cluster_manager_list_clusters_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_ListClusters_async", "segments": [ { @@ -1018,18 +2109,62 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_list_clusters_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.list_clusters", "method": { + "fullName": "google.container.v1.ClusterManager.ListClusters", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListClusters" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.ListClustersRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.ListClustersResponse", + "shortName": "list_clusters" }, + "description": "Sample for ListClusters", "file": "container_v1_generated_cluster_manager_list_clusters_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_ListClusters_sync", "segments": [ { @@ -1062,19 +2197,67 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_list_clusters_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.list_node_pools", "method": { + "fullName": "google.container.v1.ClusterManager.ListNodePools", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListNodePools" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.ListNodePoolsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.ListNodePoolsResponse", + "shortName": "list_node_pools" }, + "description": "Sample for ListNodePools", "file": "container_v1_generated_cluster_manager_list_node_pools_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_ListNodePools_async", "segments": [ { @@ -1107,18 +2290,66 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_list_node_pools_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.list_node_pools", "method": { + "fullName": "google.container.v1.ClusterManager.ListNodePools", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListNodePools" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.ListNodePoolsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.ListNodePoolsResponse", + "shortName": "list_node_pools" }, + "description": "Sample for ListNodePools", "file": "container_v1_generated_cluster_manager_list_node_pools_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_ListNodePools_sync", "segments": [ { @@ -1151,19 +2382,59 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_list_node_pools_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.list_operations", "method": { + "fullName": "google.container.v1.ClusterManager.ListOperations", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListOperations" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.ListOperationsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.ListOperationsResponse", + "shortName": "list_operations" }, + "description": "Sample for ListOperations", "file": "container_v1_generated_cluster_manager_list_operations_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_ListOperations_async", "segments": [ { @@ -1196,18 +2467,58 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_list_operations_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.list_operations", "method": { + "fullName": "google.container.v1.ClusterManager.ListOperations", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListOperations" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.ListOperationsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.ListOperationsResponse", + "shortName": "list_operations" }, + "description": "Sample for ListOperations", "file": "container_v1_generated_cluster_manager_list_operations_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_ListOperations_sync", "segments": [ { @@ -1240,19 +2551,51 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_list_operations_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.list_usable_subnetworks", "method": { + "fullName": "google.container.v1.ClusterManager.ListUsableSubnetworks", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListUsableSubnetworks" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.ListUsableSubnetworksRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.services.cluster_manager.pagers.ListUsableSubnetworksAsyncPager", + "shortName": "list_usable_subnetworks" }, + "description": "Sample for ListUsableSubnetworks", "file": "container_v1_generated_cluster_manager_list_usable_subnetworks_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_ListUsableSubnetworks_async", "segments": [ { @@ -1285,18 +2628,50 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_list_usable_subnetworks_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.list_usable_subnetworks", "method": { + "fullName": "google.container.v1.ClusterManager.ListUsableSubnetworks", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListUsableSubnetworks" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.ListUsableSubnetworksRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.services.cluster_manager.pagers.ListUsableSubnetworksPager", + "shortName": "list_usable_subnetworks" }, + "description": "Sample for ListUsableSubnetworks", "file": "container_v1_generated_cluster_manager_list_usable_subnetworks_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_ListUsableSubnetworks_sync", "segments": [ { @@ -1329,19 +2704,71 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_list_usable_subnetworks_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.rollback_node_pool_upgrade", "method": { + "fullName": "google.container.v1.ClusterManager.RollbackNodePoolUpgrade", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "RollbackNodePoolUpgrade" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.RollbackNodePoolUpgradeRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "rollback_node_pool_upgrade" }, + "description": "Sample for RollbackNodePoolUpgrade", "file": "container_v1_generated_cluster_manager_rollback_node_pool_upgrade_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_RollbackNodePoolUpgrade_async", "segments": [ { @@ -1374,18 +2801,70 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_rollback_node_pool_upgrade_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.rollback_node_pool_upgrade", "method": { + "fullName": "google.container.v1.ClusterManager.RollbackNodePoolUpgrade", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "RollbackNodePoolUpgrade" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.RollbackNodePoolUpgradeRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "rollback_node_pool_upgrade" }, + "description": "Sample for RollbackNodePoolUpgrade", "file": "container_v1_generated_cluster_manager_rollback_node_pool_upgrade_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_RollbackNodePoolUpgrade_sync", "segments": [ { @@ -1418,19 +2897,71 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_rollback_node_pool_upgrade_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_addons_config", "method": { + "fullName": "google.container.v1.ClusterManager.SetAddonsConfig", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetAddonsConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetAddonsConfigRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "addons_config", + "type": "google.cloud.container_v1.types.AddonsConfig" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_addons_config" }, + "description": "Sample for SetAddonsConfig", "file": "container_v1_generated_cluster_manager_set_addons_config_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetAddonsConfig_async", "segments": [ { @@ -1463,18 +2994,70 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_addons_config_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_addons_config", "method": { + "fullName": "google.container.v1.ClusterManager.SetAddonsConfig", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetAddonsConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetAddonsConfigRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "addons_config", + "type": "google.cloud.container_v1.types.AddonsConfig" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_addons_config" }, + "description": "Sample for SetAddonsConfig", "file": "container_v1_generated_cluster_manager_set_addons_config_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetAddonsConfig_sync", "segments": [ { @@ -1507,19 +3090,51 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_addons_config_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_labels", "method": { + "fullName": "google.container.v1.ClusterManager.SetLabels", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLabels" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetLabelsRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_labels" }, + "description": "Sample for SetLabels", "file": "container_v1_generated_cluster_manager_set_labels_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetLabels_async", "segments": [ { @@ -1552,18 +3167,50 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_labels_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_labels", "method": { + "fullName": "google.container.v1.ClusterManager.SetLabels", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLabels" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetLabelsRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_labels" }, + "description": "Sample for SetLabels", "file": "container_v1_generated_cluster_manager_set_labels_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetLabels_sync", "segments": [ { @@ -1596,19 +3243,71 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_labels_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_legacy_abac", "method": { + "fullName": "google.container.v1.ClusterManager.SetLegacyAbac", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLegacyAbac" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetLegacyAbacRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "enabled", + "type": "bool" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_legacy_abac" }, + "description": "Sample for SetLegacyAbac", "file": "container_v1_generated_cluster_manager_set_legacy_abac_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetLegacyAbac_async", "segments": [ { @@ -1641,18 +3340,70 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_legacy_abac_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_legacy_abac", "method": { + "fullName": "google.container.v1.ClusterManager.SetLegacyAbac", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLegacyAbac" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetLegacyAbacRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "enabled", + "type": "bool" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_legacy_abac" }, + "description": "Sample for SetLegacyAbac", "file": "container_v1_generated_cluster_manager_set_legacy_abac_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetLegacyAbac_sync", "segments": [ { @@ -1685,19 +3436,71 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_legacy_abac_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_locations", "method": { + "fullName": "google.container.v1.ClusterManager.SetLocations", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLocations" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetLocationsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "locations", + "type": "Sequence[str]" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_locations" }, + "description": "Sample for SetLocations", "file": "container_v1_generated_cluster_manager_set_locations_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetLocations_async", "segments": [ { @@ -1730,18 +3533,70 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_locations_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_locations", "method": { + "fullName": "google.container.v1.ClusterManager.SetLocations", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLocations" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetLocationsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "locations", + "type": "Sequence[str]" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_locations" }, + "description": "Sample for SetLocations", "file": "container_v1_generated_cluster_manager_set_locations_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetLocations_sync", "segments": [ { @@ -1774,19 +3629,71 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_locations_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_logging_service", "method": { + "fullName": "google.container.v1.ClusterManager.SetLoggingService", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLoggingService" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetLoggingServiceRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "logging_service", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_logging_service" }, + "description": "Sample for SetLoggingService", "file": "container_v1_generated_cluster_manager_set_logging_service_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetLoggingService_async", "segments": [ { @@ -1819,18 +3726,70 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_logging_service_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_logging_service", "method": { + "fullName": "google.container.v1.ClusterManager.SetLoggingService", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLoggingService" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetLoggingServiceRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "logging_service", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_logging_service" }, + "description": "Sample for SetLoggingService", "file": "container_v1_generated_cluster_manager_set_logging_service_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetLoggingService_sync", "segments": [ { @@ -1863,19 +3822,71 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_logging_service_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_maintenance_policy", "method": { + "fullName": "google.container.v1.ClusterManager.SetMaintenancePolicy", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMaintenancePolicy" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetMaintenancePolicyRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "maintenance_policy", + "type": "google.cloud.container_v1.types.MaintenancePolicy" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_maintenance_policy" }, + "description": "Sample for SetMaintenancePolicy", "file": "container_v1_generated_cluster_manager_set_maintenance_policy_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetMaintenancePolicy_async", "segments": [ { @@ -1908,18 +3919,70 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_maintenance_policy_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_maintenance_policy", "method": { + "fullName": "google.container.v1.ClusterManager.SetMaintenancePolicy", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMaintenancePolicy" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetMaintenancePolicyRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "maintenance_policy", + "type": "google.cloud.container_v1.types.MaintenancePolicy" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_maintenance_policy" }, + "description": "Sample for SetMaintenancePolicy", "file": "container_v1_generated_cluster_manager_set_maintenance_policy_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetMaintenancePolicy_sync", "segments": [ { @@ -1952,19 +4015,51 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_maintenance_policy_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_master_auth", "method": { + "fullName": "google.container.v1.ClusterManager.SetMasterAuth", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMasterAuth" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetMasterAuthRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_master_auth" }, + "description": "Sample for SetMasterAuth", "file": "container_v1_generated_cluster_manager_set_master_auth_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetMasterAuth_async", "segments": [ { @@ -1997,18 +4092,50 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_master_auth_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_master_auth", "method": { + "fullName": "google.container.v1.ClusterManager.SetMasterAuth", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMasterAuth" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetMasterAuthRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_master_auth" }, + "description": "Sample for SetMasterAuth", "file": "container_v1_generated_cluster_manager_set_master_auth_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetMasterAuth_sync", "segments": [ { @@ -2041,19 +4168,71 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_master_auth_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_monitoring_service", "method": { + "fullName": "google.container.v1.ClusterManager.SetMonitoringService", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMonitoringService" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetMonitoringServiceRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "monitoring_service", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_monitoring_service" }, + "description": "Sample for SetMonitoringService", "file": "container_v1_generated_cluster_manager_set_monitoring_service_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetMonitoringService_async", "segments": [ { @@ -2086,18 +4265,70 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_monitoring_service_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_monitoring_service", "method": { + "fullName": "google.container.v1.ClusterManager.SetMonitoringService", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMonitoringService" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetMonitoringServiceRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "monitoring_service", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_monitoring_service" }, + "description": "Sample for SetMonitoringService", "file": "container_v1_generated_cluster_manager_set_monitoring_service_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetMonitoringService_sync", "segments": [ { @@ -2130,19 +4361,71 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_monitoring_service_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_network_policy", "method": { + "fullName": "google.container.v1.ClusterManager.SetNetworkPolicy", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNetworkPolicy" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetNetworkPolicyRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "network_policy", + "type": "google.cloud.container_v1.types.NetworkPolicy" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_network_policy" }, + "description": "Sample for SetNetworkPolicy", "file": "container_v1_generated_cluster_manager_set_network_policy_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetNetworkPolicy_async", "segments": [ { @@ -2175,18 +4458,70 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_network_policy_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_network_policy", "method": { + "fullName": "google.container.v1.ClusterManager.SetNetworkPolicy", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNetworkPolicy" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetNetworkPolicyRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "network_policy", + "type": "google.cloud.container_v1.types.NetworkPolicy" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_network_policy" }, + "description": "Sample for SetNetworkPolicy", "file": "container_v1_generated_cluster_manager_set_network_policy_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetNetworkPolicy_sync", "segments": [ { @@ -2219,19 +4554,51 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_network_policy_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_node_pool_autoscaling", "method": { + "fullName": "google.container.v1.ClusterManager.SetNodePoolAutoscaling", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolAutoscaling" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetNodePoolAutoscalingRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_node_pool_autoscaling" }, + "description": "Sample for SetNodePoolAutoscaling", "file": "container_v1_generated_cluster_manager_set_node_pool_autoscaling_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetNodePoolAutoscaling_async", "segments": [ { @@ -2264,18 +4631,50 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_node_pool_autoscaling_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_node_pool_autoscaling", "method": { + "fullName": "google.container.v1.ClusterManager.SetNodePoolAutoscaling", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolAutoscaling" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetNodePoolAutoscalingRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_node_pool_autoscaling" }, + "description": "Sample for SetNodePoolAutoscaling", "file": "container_v1_generated_cluster_manager_set_node_pool_autoscaling_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetNodePoolAutoscaling_sync", "segments": [ { @@ -2308,19 +4707,51 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_node_pool_autoscaling_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_node_pool_management", "method": { + "fullName": "google.container.v1.ClusterManager.SetNodePoolManagement", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolManagement" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetNodePoolManagementRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_node_pool_management" }, + "description": "Sample for SetNodePoolManagement", "file": "container_v1_generated_cluster_manager_set_node_pool_management_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetNodePoolManagement_async", "segments": [ { @@ -2353,18 +4784,50 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_node_pool_management_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_node_pool_management", "method": { + "fullName": "google.container.v1.ClusterManager.SetNodePoolManagement", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolManagement" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetNodePoolManagementRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_node_pool_management" }, + "description": "Sample for SetNodePoolManagement", "file": "container_v1_generated_cluster_manager_set_node_pool_management_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetNodePoolManagement_sync", "segments": [ { @@ -2397,19 +4860,51 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_node_pool_management_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.set_node_pool_size", "method": { + "fullName": "google.container.v1.ClusterManager.SetNodePoolSize", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolSize" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetNodePoolSizeRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_node_pool_size" }, + "description": "Sample for SetNodePoolSize", "file": "container_v1_generated_cluster_manager_set_node_pool_size_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetNodePoolSize_async", "segments": [ { @@ -2442,18 +4937,50 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_node_pool_size_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.set_node_pool_size", "method": { + "fullName": "google.container.v1.ClusterManager.SetNodePoolSize", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolSize" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.SetNodePoolSizeRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "set_node_pool_size" }, + "description": "Sample for SetNodePoolSize", "file": "container_v1_generated_cluster_manager_set_node_pool_size_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_SetNodePoolSize_sync", "segments": [ { @@ -2486,19 +5013,67 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_set_node_pool_size_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.start_ip_rotation", "method": { + "fullName": "google.container.v1.ClusterManager.StartIPRotation", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "StartIPRotation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.StartIPRotationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "start_ip_rotation" }, + "description": "Sample for StartIPRotation", "file": "container_v1_generated_cluster_manager_start_ip_rotation_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_StartIPRotation_async", "segments": [ { @@ -2531,18 +5106,66 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_start_ip_rotation_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.start_ip_rotation", "method": { + "fullName": "google.container.v1.ClusterManager.StartIPRotation", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "StartIPRotation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.StartIPRotationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "start_ip_rotation" }, + "description": "Sample for StartIPRotation", "file": "container_v1_generated_cluster_manager_start_ip_rotation_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_StartIPRotation_sync", "segments": [ { @@ -2575,19 +5198,71 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_start_ip_rotation_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.update_cluster", "method": { + "fullName": "google.container.v1.ClusterManager.UpdateCluster", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.UpdateClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "update", + "type": "google.cloud.container_v1.types.ClusterUpdate" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "update_cluster" }, + "description": "Sample for UpdateCluster", "file": "container_v1_generated_cluster_manager_update_cluster_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_UpdateCluster_async", "segments": [ { @@ -2620,18 +5295,70 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_update_cluster_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.update_cluster", "method": { + "fullName": "google.container.v1.ClusterManager.UpdateCluster", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.UpdateClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "update", + "type": "google.cloud.container_v1.types.ClusterUpdate" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "update_cluster" }, + "description": "Sample for UpdateCluster", "file": "container_v1_generated_cluster_manager_update_cluster_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_UpdateCluster_sync", "segments": [ { @@ -2664,19 +5391,71 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_update_cluster_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.update_master", "method": { + "fullName": "google.container.v1.ClusterManager.UpdateMaster", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateMaster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.UpdateMasterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "master_version", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "update_master" }, + "description": "Sample for UpdateMaster", "file": "container_v1_generated_cluster_manager_update_master_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_UpdateMaster_async", "segments": [ { @@ -2709,18 +5488,70 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_update_master_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.update_master", "method": { + "fullName": "google.container.v1.ClusterManager.UpdateMaster", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateMaster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.UpdateMasterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "master_version", + "type": "str" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "update_master" }, + "description": "Sample for UpdateMaster", "file": "container_v1_generated_cluster_manager_update_master_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_UpdateMaster_sync", "segments": [ { @@ -2753,19 +5584,51 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_update_master_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerAsyncClient.update_node_pool", "method": { + "fullName": "google.container.v1.ClusterManager.UpdateNodePool", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.UpdateNodePoolRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "update_node_pool" }, + "description": "Sample for UpdateNodePool", "file": "container_v1_generated_cluster_manager_update_node_pool_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_UpdateNodePool_async", "segments": [ { @@ -2798,18 +5661,50 @@ "start": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_update_node_pool_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1.ClusterManagerClient.update_node_pool", "method": { + "fullName": "google.container.v1.ClusterManager.UpdateNodePool", "service": { + "fullName": "google.container.v1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1.types.UpdateNodePoolRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1.types.Operation", + "shortName": "update_node_pool" }, + "description": "Sample for UpdateNodePool", "file": "container_v1_generated_cluster_manager_update_node_pool_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1_generated_ClusterManager_UpdateNodePool_sync", "segments": [ { @@ -2842,7 +5737,8 @@ "start": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1_generated_cluster_manager_update_node_pool_sync.py" } ] } diff --git a/samples/generated_samples/snippet_metadata_container_v1beta1.json b/samples/generated_samples/snippet_metadata_container_v1beta1.json index 82fb9531..850ebf58 100644 --- a/samples/generated_samples/snippet_metadata_container_v1beta1.json +++ b/samples/generated_samples/snippet_metadata_container_v1beta1.json @@ -1,16 +1,68 @@ { + "clientLibrary": { + "apis": [ + { + "id": "google.container.v1beta1", + "version": "v1beta1" + } + ], + "language": "PYTHON", + "name": "google-cloud-container" + }, "snippets": [ { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.cancel_operation", "method": { + "fullName": "google.container.v1beta1.ClusterManager.CancelOperation", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CancelOperation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.CancelOperationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "operation_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "shortName": "cancel_operation" }, + "description": "Sample for CancelOperation", "file": "container_v1beta1_generated_cluster_manager_cancel_operation_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_CancelOperation_async", "segments": [ { @@ -41,18 +93,61 @@ "end": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_cancel_operation_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.cancel_operation", "method": { + "fullName": "google.container.v1beta1.ClusterManager.CancelOperation", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CancelOperation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.CancelOperationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "operation_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "shortName": "cancel_operation" }, + "description": "Sample for CancelOperation", "file": "container_v1beta1_generated_cluster_manager_cancel_operation_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_CancelOperation_sync", "segments": [ { @@ -83,19 +178,63 @@ "end": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_cancel_operation_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.complete_ip_rotation", "method": { + "fullName": "google.container.v1beta1.ClusterManager.CompleteIPRotation", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CompleteIPRotation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.CompleteIPRotationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "complete_ip_rotation" }, + "description": "Sample for CompleteIPRotation", "file": "container_v1beta1_generated_cluster_manager_complete_ip_rotation_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_CompleteIPRotation_async", "segments": [ { @@ -128,18 +267,62 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_complete_ip_rotation_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.complete_ip_rotation", "method": { + "fullName": "google.container.v1beta1.ClusterManager.CompleteIPRotation", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CompleteIPRotation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.CompleteIPRotationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "complete_ip_rotation" }, + "description": "Sample for CompleteIPRotation", "file": "container_v1beta1_generated_cluster_manager_complete_ip_rotation_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_CompleteIPRotation_sync", "segments": [ { @@ -172,19 +355,63 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_complete_ip_rotation_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.create_cluster", "method": { + "fullName": "google.container.v1beta1.ClusterManager.CreateCluster", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CreateCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.CreateClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster", + "type": "google.cloud.container_v1beta1.types.Cluster" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "create_cluster" }, + "description": "Sample for CreateCluster", "file": "container_v1beta1_generated_cluster_manager_create_cluster_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_CreateCluster_async", "segments": [ { @@ -217,18 +444,62 @@ "start": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_create_cluster_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.create_cluster", "method": { + "fullName": "google.container.v1beta1.ClusterManager.CreateCluster", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CreateCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.CreateClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster", + "type": "google.cloud.container_v1beta1.types.Cluster" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "create_cluster" }, + "description": "Sample for CreateCluster", "file": "container_v1beta1_generated_cluster_manager_create_cluster_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_CreateCluster_sync", "segments": [ { @@ -261,19 +532,67 @@ "start": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_create_cluster_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.create_node_pool", "method": { + "fullName": "google.container.v1beta1.ClusterManager.CreateNodePool", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CreateNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.CreateNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool", + "type": "google.cloud.container_v1beta1.types.NodePool" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "create_node_pool" }, + "description": "Sample for CreateNodePool", "file": "container_v1beta1_generated_cluster_manager_create_node_pool_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_CreateNodePool_async", "segments": [ { @@ -306,18 +625,66 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_create_node_pool_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.create_node_pool", "method": { + "fullName": "google.container.v1beta1.ClusterManager.CreateNodePool", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "CreateNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.CreateNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool", + "type": "google.cloud.container_v1beta1.types.NodePool" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "create_node_pool" }, + "description": "Sample for CreateNodePool", "file": "container_v1beta1_generated_cluster_manager_create_node_pool_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_CreateNodePool_sync", "segments": [ { @@ -350,19 +717,63 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_create_node_pool_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.delete_cluster", "method": { + "fullName": "google.container.v1beta1.ClusterManager.DeleteCluster", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "DeleteCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.DeleteClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "delete_cluster" }, + "description": "Sample for DeleteCluster", "file": "container_v1beta1_generated_cluster_manager_delete_cluster_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_DeleteCluster_async", "segments": [ { @@ -395,18 +806,62 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_delete_cluster_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.delete_cluster", "method": { + "fullName": "google.container.v1beta1.ClusterManager.DeleteCluster", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "DeleteCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.DeleteClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "delete_cluster" }, + "description": "Sample for DeleteCluster", "file": "container_v1beta1_generated_cluster_manager_delete_cluster_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_DeleteCluster_sync", "segments": [ { @@ -439,19 +894,67 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_delete_cluster_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.delete_node_pool", "method": { + "fullName": "google.container.v1beta1.ClusterManager.DeleteNodePool", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "DeleteNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.DeleteNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "delete_node_pool" }, + "description": "Sample for DeleteNodePool", "file": "container_v1beta1_generated_cluster_manager_delete_node_pool_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_DeleteNodePool_async", "segments": [ { @@ -484,18 +987,66 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_delete_node_pool_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.delete_node_pool", "method": { + "fullName": "google.container.v1beta1.ClusterManager.DeleteNodePool", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "DeleteNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.DeleteNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "delete_node_pool" }, + "description": "Sample for DeleteNodePool", "file": "container_v1beta1_generated_cluster_manager_delete_node_pool_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_DeleteNodePool_sync", "segments": [ { @@ -528,19 +1079,63 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_delete_node_pool_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.get_cluster", "method": { + "fullName": "google.container.v1beta1.ClusterManager.GetCluster", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.GetClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Cluster", + "shortName": "get_cluster" }, + "description": "Sample for GetCluster", "file": "container_v1beta1_generated_cluster_manager_get_cluster_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_GetCluster_async", "segments": [ { @@ -573,18 +1168,62 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_get_cluster_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.get_cluster", "method": { + "fullName": "google.container.v1beta1.ClusterManager.GetCluster", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.GetClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Cluster", + "shortName": "get_cluster" }, + "description": "Sample for GetCluster", "file": "container_v1beta1_generated_cluster_manager_get_cluster_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_GetCluster_sync", "segments": [ { @@ -617,19 +1256,51 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_get_cluster_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.get_json_web_keys", "method": { + "fullName": "google.container.v1beta1.ClusterManager.GetJSONWebKeys", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetJSONWebKeys" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.GetJSONWebKeysRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.GetJSONWebKeysResponse", + "shortName": "get_json_web_keys" }, + "description": "Sample for GetJSONWebKeys", "file": "container_v1beta1_generated_cluster_manager_get_json_web_keys_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_GetJSONWebKeys_async", "segments": [ { @@ -662,18 +1333,50 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_get_json_web_keys_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.get_json_web_keys", "method": { + "fullName": "google.container.v1beta1.ClusterManager.GetJSONWebKeys", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetJSONWebKeys" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.GetJSONWebKeysRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.GetJSONWebKeysResponse", + "shortName": "get_json_web_keys" }, + "description": "Sample for GetJSONWebKeys", "file": "container_v1beta1_generated_cluster_manager_get_json_web_keys_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_GetJSONWebKeys_sync", "segments": [ { @@ -706,19 +1409,67 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_get_json_web_keys_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.get_node_pool", "method": { + "fullName": "google.container.v1beta1.ClusterManager.GetNodePool", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.GetNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.NodePool", + "shortName": "get_node_pool" }, + "description": "Sample for GetNodePool", "file": "container_v1beta1_generated_cluster_manager_get_node_pool_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_GetNodePool_async", "segments": [ { @@ -751,18 +1502,66 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_get_node_pool_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.get_node_pool", "method": { + "fullName": "google.container.v1beta1.ClusterManager.GetNodePool", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.GetNodePoolRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.NodePool", + "shortName": "get_node_pool" }, + "description": "Sample for GetNodePool", "file": "container_v1beta1_generated_cluster_manager_get_node_pool_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_GetNodePool_sync", "segments": [ { @@ -795,19 +1594,63 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_get_node_pool_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.get_operation", "method": { + "fullName": "google.container.v1beta1.ClusterManager.GetOperation", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetOperation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.GetOperationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "operation_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "get_operation" }, + "description": "Sample for GetOperation", "file": "container_v1beta1_generated_cluster_manager_get_operation_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_GetOperation_async", "segments": [ { @@ -840,18 +1683,62 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_get_operation_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.get_operation", "method": { + "fullName": "google.container.v1beta1.ClusterManager.GetOperation", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetOperation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.GetOperationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "operation_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "get_operation" }, + "description": "Sample for GetOperation", "file": "container_v1beta1_generated_cluster_manager_get_operation_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_GetOperation_sync", "segments": [ { @@ -884,19 +1771,59 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_get_operation_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.get_server_config", "method": { + "fullName": "google.container.v1beta1.ClusterManager.GetServerConfig", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetServerConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.GetServerConfigRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.ServerConfig", + "shortName": "get_server_config" }, + "description": "Sample for GetServerConfig", "file": "container_v1beta1_generated_cluster_manager_get_server_config_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_GetServerConfig_async", "segments": [ { @@ -929,18 +1856,58 @@ "start": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_get_server_config_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.get_server_config", "method": { + "fullName": "google.container.v1beta1.ClusterManager.GetServerConfig", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "GetServerConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.GetServerConfigRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.ServerConfig", + "shortName": "get_server_config" }, + "description": "Sample for GetServerConfig", "file": "container_v1beta1_generated_cluster_manager_get_server_config_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_GetServerConfig_sync", "segments": [ { @@ -973,19 +1940,59 @@ "start": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_get_server_config_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.list_clusters", "method": { + "fullName": "google.container.v1beta1.ClusterManager.ListClusters", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListClusters" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.ListClustersRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.ListClustersResponse", + "shortName": "list_clusters" }, + "description": "Sample for ListClusters", "file": "container_v1beta1_generated_cluster_manager_list_clusters_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_ListClusters_async", "segments": [ { @@ -1018,18 +2025,58 @@ "start": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_list_clusters_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.list_clusters", "method": { + "fullName": "google.container.v1beta1.ClusterManager.ListClusters", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListClusters" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.ListClustersRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.ListClustersResponse", + "shortName": "list_clusters" }, + "description": "Sample for ListClusters", "file": "container_v1beta1_generated_cluster_manager_list_clusters_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_ListClusters_sync", "segments": [ { @@ -1062,19 +2109,55 @@ "start": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_list_clusters_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.list_locations", "method": { + "fullName": "google.container.v1beta1.ClusterManager.ListLocations", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListLocations" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.ListLocationsRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.ListLocationsResponse", + "shortName": "list_locations" }, + "description": "Sample for ListLocations", "file": "container_v1beta1_generated_cluster_manager_list_locations_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_ListLocations_async", "segments": [ { @@ -1107,18 +2190,54 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_list_locations_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.list_locations", "method": { + "fullName": "google.container.v1beta1.ClusterManager.ListLocations", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListLocations" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.ListLocationsRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.ListLocationsResponse", + "shortName": "list_locations" }, + "description": "Sample for ListLocations", "file": "container_v1beta1_generated_cluster_manager_list_locations_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_ListLocations_sync", "segments": [ { @@ -1151,19 +2270,63 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_list_locations_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.list_node_pools", "method": { + "fullName": "google.container.v1beta1.ClusterManager.ListNodePools", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListNodePools" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.ListNodePoolsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.ListNodePoolsResponse", + "shortName": "list_node_pools" }, + "description": "Sample for ListNodePools", "file": "container_v1beta1_generated_cluster_manager_list_node_pools_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_ListNodePools_async", "segments": [ { @@ -1196,18 +2359,62 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_list_node_pools_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.list_node_pools", "method": { + "fullName": "google.container.v1beta1.ClusterManager.ListNodePools", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListNodePools" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.ListNodePoolsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.ListNodePoolsResponse", + "shortName": "list_node_pools" }, + "description": "Sample for ListNodePools", "file": "container_v1beta1_generated_cluster_manager_list_node_pools_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_ListNodePools_sync", "segments": [ { @@ -1240,19 +2447,59 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_list_node_pools_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.list_operations", "method": { + "fullName": "google.container.v1beta1.ClusterManager.ListOperations", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListOperations" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.ListOperationsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.ListOperationsResponse", + "shortName": "list_operations" }, + "description": "Sample for ListOperations", "file": "container_v1beta1_generated_cluster_manager_list_operations_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_ListOperations_async", "segments": [ { @@ -1285,18 +2532,58 @@ "start": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_list_operations_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.list_operations", "method": { + "fullName": "google.container.v1beta1.ClusterManager.ListOperations", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListOperations" - } - }, - "file": "container_v1beta1_generated_cluster_manager_list_operations_sync.py", + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.ListOperationsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.ListOperationsResponse", + "shortName": "list_operations" + }, + "description": "Sample for ListOperations", + "file": "container_v1beta1_generated_cluster_manager_list_operations_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_ListOperations_sync", "segments": [ { @@ -1329,19 +2616,55 @@ "start": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_list_operations_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.list_usable_subnetworks", "method": { + "fullName": "google.container.v1beta1.ClusterManager.ListUsableSubnetworks", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListUsableSubnetworks" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.ListUsableSubnetworksRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.services.cluster_manager.pagers.ListUsableSubnetworksAsyncPager", + "shortName": "list_usable_subnetworks" }, + "description": "Sample for ListUsableSubnetworks", "file": "container_v1beta1_generated_cluster_manager_list_usable_subnetworks_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_ListUsableSubnetworks_async", "segments": [ { @@ -1374,18 +2697,54 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_list_usable_subnetworks_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.list_usable_subnetworks", "method": { + "fullName": "google.container.v1beta1.ClusterManager.ListUsableSubnetworks", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "ListUsableSubnetworks" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.ListUsableSubnetworksRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.services.cluster_manager.pagers.ListUsableSubnetworksPager", + "shortName": "list_usable_subnetworks" }, + "description": "Sample for ListUsableSubnetworks", "file": "container_v1beta1_generated_cluster_manager_list_usable_subnetworks_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_ListUsableSubnetworks_sync", "segments": [ { @@ -1418,19 +2777,67 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_list_usable_subnetworks_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.rollback_node_pool_upgrade", "method": { + "fullName": "google.container.v1beta1.ClusterManager.RollbackNodePoolUpgrade", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "RollbackNodePoolUpgrade" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.RollbackNodePoolUpgradeRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "rollback_node_pool_upgrade" }, + "description": "Sample for RollbackNodePoolUpgrade", "file": "container_v1beta1_generated_cluster_manager_rollback_node_pool_upgrade_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_RollbackNodePoolUpgrade_async", "segments": [ { @@ -1463,18 +2870,66 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_rollback_node_pool_upgrade_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.rollback_node_pool_upgrade", "method": { + "fullName": "google.container.v1beta1.ClusterManager.RollbackNodePoolUpgrade", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "RollbackNodePoolUpgrade" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.RollbackNodePoolUpgradeRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "rollback_node_pool_upgrade" }, + "description": "Sample for RollbackNodePoolUpgrade", "file": "container_v1beta1_generated_cluster_manager_rollback_node_pool_upgrade_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_RollbackNodePoolUpgrade_sync", "segments": [ { @@ -1507,19 +2962,67 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_rollback_node_pool_upgrade_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_addons_config", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetAddonsConfig", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetAddonsConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetAddonsConfigRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "addons_config", + "type": "google.cloud.container_v1beta1.types.AddonsConfig" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_addons_config" }, + "description": "Sample for SetAddonsConfig", "file": "container_v1beta1_generated_cluster_manager_set_addons_config_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetAddonsConfig_async", "segments": [ { @@ -1552,18 +3055,66 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_addons_config_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_addons_config", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetAddonsConfig", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetAddonsConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetAddonsConfigRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "addons_config", + "type": "google.cloud.container_v1beta1.types.AddonsConfig" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_addons_config" }, + "description": "Sample for SetAddonsConfig", "file": "container_v1beta1_generated_cluster_manager_set_addons_config_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetAddonsConfig_sync", "segments": [ { @@ -1596,19 +3147,71 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_addons_config_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_labels", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetLabels", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLabels" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetLabelsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "resource_labels", + "type": "Mapping[str, str]" + }, + { + "name": "label_fingerprint", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_labels" }, + "description": "Sample for SetLabels", "file": "container_v1beta1_generated_cluster_manager_set_labels_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetLabels_async", "segments": [ { @@ -1641,18 +3244,70 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_labels_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_labels", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetLabels", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLabels" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetLabelsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "resource_labels", + "type": "Mapping[str, str]" + }, + { + "name": "label_fingerprint", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_labels" }, + "description": "Sample for SetLabels", "file": "container_v1beta1_generated_cluster_manager_set_labels_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetLabels_sync", "segments": [ { @@ -1685,19 +3340,67 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_labels_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_legacy_abac", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetLegacyAbac", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLegacyAbac" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetLegacyAbacRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "enabled", + "type": "bool" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_legacy_abac" }, + "description": "Sample for SetLegacyAbac", "file": "container_v1beta1_generated_cluster_manager_set_legacy_abac_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetLegacyAbac_async", "segments": [ { @@ -1730,18 +3433,66 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_legacy_abac_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_legacy_abac", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetLegacyAbac", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLegacyAbac" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetLegacyAbacRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "enabled", + "type": "bool" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_legacy_abac" }, + "description": "Sample for SetLegacyAbac", "file": "container_v1beta1_generated_cluster_manager_set_legacy_abac_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetLegacyAbac_sync", "segments": [ { @@ -1774,19 +3525,67 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_legacy_abac_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_locations", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetLocations", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLocations" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetLocationsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "locations", + "type": "Sequence[str]" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_locations" }, + "description": "Sample for SetLocations", "file": "container_v1beta1_generated_cluster_manager_set_locations_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetLocations_async", "segments": [ { @@ -1819,18 +3618,66 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_locations_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_locations", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetLocations", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLocations" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetLocationsRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "locations", + "type": "Sequence[str]" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_locations" }, + "description": "Sample for SetLocations", "file": "container_v1beta1_generated_cluster_manager_set_locations_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetLocations_sync", "segments": [ { @@ -1863,19 +3710,67 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_locations_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_logging_service", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetLoggingService", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLoggingService" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetLoggingServiceRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "logging_service", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_logging_service" }, + "description": "Sample for SetLoggingService", "file": "container_v1beta1_generated_cluster_manager_set_logging_service_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetLoggingService_async", "segments": [ { @@ -1908,18 +3803,66 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_logging_service_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_logging_service", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetLoggingService", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetLoggingService" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetLoggingServiceRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "logging_service", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_logging_service" }, + "description": "Sample for SetLoggingService", "file": "container_v1beta1_generated_cluster_manager_set_logging_service_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetLoggingService_sync", "segments": [ { @@ -1952,19 +3895,67 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_logging_service_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_maintenance_policy", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetMaintenancePolicy", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMaintenancePolicy" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetMaintenancePolicyRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "maintenance_policy", + "type": "google.cloud.container_v1beta1.types.MaintenancePolicy" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_maintenance_policy" }, + "description": "Sample for SetMaintenancePolicy", "file": "container_v1beta1_generated_cluster_manager_set_maintenance_policy_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetMaintenancePolicy_async", "segments": [ { @@ -1997,18 +3988,66 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_maintenance_policy_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_maintenance_policy", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetMaintenancePolicy", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMaintenancePolicy" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetMaintenancePolicyRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "maintenance_policy", + "type": "google.cloud.container_v1beta1.types.MaintenancePolicy" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_maintenance_policy" }, + "description": "Sample for SetMaintenancePolicy", "file": "container_v1beta1_generated_cluster_manager_set_maintenance_policy_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetMaintenancePolicy_sync", "segments": [ { @@ -2041,19 +4080,51 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_maintenance_policy_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_master_auth", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetMasterAuth", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMasterAuth" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetMasterAuthRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_master_auth" }, + "description": "Sample for SetMasterAuth", "file": "container_v1beta1_generated_cluster_manager_set_master_auth_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetMasterAuth_async", "segments": [ { @@ -2086,18 +4157,50 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_master_auth_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_master_auth", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetMasterAuth", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMasterAuth" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetMasterAuthRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_master_auth" }, + "description": "Sample for SetMasterAuth", "file": "container_v1beta1_generated_cluster_manager_set_master_auth_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetMasterAuth_sync", "segments": [ { @@ -2130,19 +4233,67 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_master_auth_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_monitoring_service", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetMonitoringService", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMonitoringService" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetMonitoringServiceRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "monitoring_service", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_monitoring_service" }, + "description": "Sample for SetMonitoringService", "file": "container_v1beta1_generated_cluster_manager_set_monitoring_service_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetMonitoringService_async", "segments": [ { @@ -2175,18 +4326,66 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_monitoring_service_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_monitoring_service", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetMonitoringService", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetMonitoringService" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetMonitoringServiceRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "monitoring_service", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_monitoring_service" }, + "description": "Sample for SetMonitoringService", "file": "container_v1beta1_generated_cluster_manager_set_monitoring_service_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetMonitoringService_sync", "segments": [ { @@ -2219,19 +4418,67 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_monitoring_service_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_network_policy", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetNetworkPolicy", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNetworkPolicy" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetNetworkPolicyRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "network_policy", + "type": "google.cloud.container_v1beta1.types.NetworkPolicy" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_network_policy" }, + "description": "Sample for SetNetworkPolicy", "file": "container_v1beta1_generated_cluster_manager_set_network_policy_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetNetworkPolicy_async", "segments": [ { @@ -2264,18 +4511,66 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_network_policy_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_network_policy", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetNetworkPolicy", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNetworkPolicy" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetNetworkPolicyRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "network_policy", + "type": "google.cloud.container_v1beta1.types.NetworkPolicy" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_network_policy" }, + "description": "Sample for SetNetworkPolicy", "file": "container_v1beta1_generated_cluster_manager_set_network_policy_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetNetworkPolicy_sync", "segments": [ { @@ -2308,19 +4603,51 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_network_policy_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_node_pool_autoscaling", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetNodePoolAutoscaling", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolAutoscaling" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetNodePoolAutoscalingRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_node_pool_autoscaling" }, + "description": "Sample for SetNodePoolAutoscaling", "file": "container_v1beta1_generated_cluster_manager_set_node_pool_autoscaling_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetNodePoolAutoscaling_async", "segments": [ { @@ -2353,18 +4680,50 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_node_pool_autoscaling_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_node_pool_autoscaling", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetNodePoolAutoscaling", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolAutoscaling" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetNodePoolAutoscalingRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_node_pool_autoscaling" }, + "description": "Sample for SetNodePoolAutoscaling", "file": "container_v1beta1_generated_cluster_manager_set_node_pool_autoscaling_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetNodePoolAutoscaling_sync", "segments": [ { @@ -2397,19 +4756,71 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_node_pool_autoscaling_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_node_pool_management", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetNodePoolManagement", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolManagement" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetNodePoolManagementRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "management", + "type": "google.cloud.container_v1beta1.types.NodeManagement" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_node_pool_management" }, + "description": "Sample for SetNodePoolManagement", "file": "container_v1beta1_generated_cluster_manager_set_node_pool_management_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetNodePoolManagement_async", "segments": [ { @@ -2442,18 +4853,70 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_node_pool_management_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_node_pool_management", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetNodePoolManagement", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolManagement" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetNodePoolManagementRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "node_pool_id", + "type": "str" + }, + { + "name": "management", + "type": "google.cloud.container_v1beta1.types.NodeManagement" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_node_pool_management" }, + "description": "Sample for SetNodePoolManagement", "file": "container_v1beta1_generated_cluster_manager_set_node_pool_management_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetNodePoolManagement_sync", "segments": [ { @@ -2486,19 +4949,51 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_node_pool_management_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.set_node_pool_size", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetNodePoolSize", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolSize" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetNodePoolSizeRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_node_pool_size" }, + "description": "Sample for SetNodePoolSize", "file": "container_v1beta1_generated_cluster_manager_set_node_pool_size_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetNodePoolSize_async", "segments": [ { @@ -2531,18 +5026,50 @@ "start": 46, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_node_pool_size_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.set_node_pool_size", "method": { + "fullName": "google.container.v1beta1.ClusterManager.SetNodePoolSize", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "SetNodePoolSize" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.SetNodePoolSizeRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "set_node_pool_size" }, + "description": "Sample for SetNodePoolSize", "file": "container_v1beta1_generated_cluster_manager_set_node_pool_size_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_SetNodePoolSize_sync", "segments": [ { @@ -2575,19 +5102,63 @@ "start": 46, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_set_node_pool_size_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.start_ip_rotation", "method": { + "fullName": "google.container.v1beta1.ClusterManager.StartIPRotation", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "StartIPRotation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.StartIPRotationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "start_ip_rotation" }, + "description": "Sample for StartIPRotation", "file": "container_v1beta1_generated_cluster_manager_start_ip_rotation_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_StartIPRotation_async", "segments": [ { @@ -2620,18 +5191,62 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_start_ip_rotation_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.start_ip_rotation", "method": { + "fullName": "google.container.v1beta1.ClusterManager.StartIPRotation", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "StartIPRotation" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.StartIPRotationRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "start_ip_rotation" }, + "description": "Sample for StartIPRotation", "file": "container_v1beta1_generated_cluster_manager_start_ip_rotation_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_StartIPRotation_sync", "segments": [ { @@ -2664,19 +5279,67 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_start_ip_rotation_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.update_cluster", "method": { + "fullName": "google.container.v1beta1.ClusterManager.UpdateCluster", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.UpdateClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "update", + "type": "google.cloud.container_v1beta1.types.ClusterUpdate" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "update_cluster" }, + "description": "Sample for UpdateCluster", "file": "container_v1beta1_generated_cluster_manager_update_cluster_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_UpdateCluster_async", "segments": [ { @@ -2709,18 +5372,66 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_update_cluster_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.update_cluster", "method": { + "fullName": "google.container.v1beta1.ClusterManager.UpdateCluster", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateCluster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.UpdateClusterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "update", + "type": "google.cloud.container_v1beta1.types.ClusterUpdate" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "update_cluster" }, + "description": "Sample for UpdateCluster", "file": "container_v1beta1_generated_cluster_manager_update_cluster_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_UpdateCluster_sync", "segments": [ { @@ -2753,19 +5464,67 @@ "start": 44, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_update_cluster_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.update_master", "method": { + "fullName": "google.container.v1beta1.ClusterManager.UpdateMaster", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateMaster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.UpdateMasterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "master_version", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "update_master" }, + "description": "Sample for UpdateMaster", "file": "container_v1beta1_generated_cluster_manager_update_master_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_UpdateMaster_async", "segments": [ { @@ -2798,18 +5557,66 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_update_master_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.update_master", "method": { + "fullName": "google.container.v1beta1.ClusterManager.UpdateMaster", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateMaster" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.UpdateMasterRequest" + }, + { + "name": "project_id", + "type": "str" + }, + { + "name": "zone", + "type": "str" + }, + { + "name": "cluster_id", + "type": "str" + }, + { + "name": "master_version", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "update_master" }, + "description": "Sample for UpdateMaster", "file": "container_v1beta1_generated_cluster_manager_update_master_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_UpdateMaster_sync", "segments": [ { @@ -2842,19 +5649,51 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_update_master_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient", + "shortName": "ClusterManagerAsyncClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerAsyncClient.update_node_pool", "method": { + "fullName": "google.container.v1beta1.ClusterManager.UpdateNodePool", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.UpdateNodePoolRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "update_node_pool" }, + "description": "Sample for UpdateNodePool", "file": "container_v1beta1_generated_cluster_manager_update_node_pool_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_UpdateNodePool_async", "segments": [ { @@ -2887,18 +5726,50 @@ "start": 47, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_update_node_pool_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient", + "shortName": "ClusterManagerClient" + }, + "fullName": "google.cloud.container_v1beta1.ClusterManagerClient.update_node_pool", "method": { + "fullName": "google.container.v1beta1.ClusterManager.UpdateNodePool", "service": { + "fullName": "google.container.v1beta1.ClusterManager", "shortName": "ClusterManager" }, "shortName": "UpdateNodePool" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.container_v1beta1.types.UpdateNodePoolRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.container_v1beta1.types.Operation", + "shortName": "update_node_pool" }, + "description": "Sample for UpdateNodePool", "file": "container_v1beta1_generated_cluster_manager_update_node_pool_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "container_v1beta1_generated_ClusterManager_UpdateNodePool_sync", "segments": [ { @@ -2931,7 +5802,8 @@ "start": 47, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "container_v1beta1_generated_cluster_manager_update_node_pool_sync.py" } ] } diff --git a/samples/snippets/create_cluster.py b/samples/snippets/create_cluster.py index ad5ddfc3..949c6f6e 100644 --- a/samples/snippets/create_cluster.py +++ b/samples/snippets/create_cluster.py @@ -97,7 +97,8 @@ def create_cluster(project_id: str, location: str, cluster_name: str) -> None: if __name__ == "__main__": parser = argparse.ArgumentParser( - description=__doc__, formatter_class=argparse.RawDescriptionHelpFormatter, + description=__doc__, + formatter_class=argparse.RawDescriptionHelpFormatter, ) parser.add_argument("project_id", help="Google Cloud project ID") parser.add_argument("zone", help="GKE Cluster zone") diff --git a/samples/snippets/create_cluster_test.py b/samples/snippets/create_cluster_test.py index 728c0750..75a281c9 100644 --- a/samples/snippets/create_cluster_test.py +++ b/samples/snippets/create_cluster_test.py @@ -17,9 +17,7 @@ import uuid import backoff - from google.cloud import container_v1 as gke - import pytest import create_cluster as gke_create diff --git a/samples/snippets/delete_cluster.py b/samples/snippets/delete_cluster.py index 0290766e..405de978 100644 --- a/samples/snippets/delete_cluster.py +++ b/samples/snippets/delete_cluster.py @@ -90,7 +90,8 @@ def delete_cluster(project_id: str, location: str, cluster_name: str) -> None: if __name__ == "__main__": parser = argparse.ArgumentParser( - description=__doc__, formatter_class=argparse.RawDescriptionHelpFormatter, + description=__doc__, + formatter_class=argparse.RawDescriptionHelpFormatter, ) parser.add_argument("project_id", help="Google Cloud project ID") parser.add_argument("zone", help="GKE Cluster zone") diff --git a/samples/snippets/delete_cluster_test.py b/samples/snippets/delete_cluster_test.py index fc7845d3..5e187d84 100644 --- a/samples/snippets/delete_cluster_test.py +++ b/samples/snippets/delete_cluster_test.py @@ -17,10 +17,8 @@ import uuid import backoff - from google.api_core import exceptions as googleEx from google.cloud import container_v1 as gke - import pytest import delete_cluster as gke_delete diff --git a/samples/snippets/noxfile.py b/samples/snippets/noxfile.py index 85f5836d..3b3ffa5d 100644 --- a/samples/snippets/noxfile.py +++ b/samples/snippets/noxfile.py @@ -22,14 +22,14 @@ import nox - # WARNING - WARNING - WARNING - WARNING - WARNING # WARNING - WARNING - WARNING - WARNING - WARNING # DO NOT EDIT THIS FILE EVER! # WARNING - WARNING - WARNING - WARNING - WARNING # WARNING - WARNING - WARNING - WARNING - WARNING -BLACK_VERSION = "black==19.10b0" +BLACK_VERSION = "black==22.3.0" +ISORT_VERSION = "isort==5.10.1" # Copy `noxfile_config.py` to your directory and modify it instead. @@ -168,12 +168,33 @@ def lint(session: nox.sessions.Session) -> None: @nox.session def blacken(session: nox.sessions.Session) -> None: + """Run black. Format code to uniform standard.""" session.install(BLACK_VERSION) python_files = [path for path in os.listdir(".") if path.endswith(".py")] session.run("black", *python_files) +# +# format = isort + black +# + + +@nox.session +def format(session: nox.sessions.Session) -> None: + """ + Run isort to sort imports. Then run black + to format code to uniform standard. + """ + session.install(BLACK_VERSION, ISORT_VERSION) + python_files = [path for path in os.listdir(".") if path.endswith(".py")] + + # Use the --fss option to sort imports using strict alphabetical order. + # See https://pycqa.github.io/isort/docs/configuration/options.html#force-sort-within-sections + session.run("isort", "--fss", *python_files) + session.run("black", *python_files) + + # # Sample Tests # @@ -253,7 +274,7 @@ def py(session: nox.sessions.Session) -> None: def _get_repo_root() -> Optional[str]: - """ Returns the root folder of the project. """ + """Returns the root folder of the project.""" # Get root of this repository. Assume we don't have directories nested deeper than 10 items. p = Path(os.getcwd()) for i in range(10): diff --git a/samples/snippets/quickstart.py b/samples/snippets/quickstart.py index b7cecb52..3299a9a1 100644 --- a/samples/snippets/quickstart.py +++ b/samples/snippets/quickstart.py @@ -43,7 +43,8 @@ def list_clusters(project_id: str, location: str) -> None: if __name__ == "__main__": parser = argparse.ArgumentParser( - description=__doc__, formatter_class=argparse.RawDescriptionHelpFormatter, + description=__doc__, + formatter_class=argparse.RawDescriptionHelpFormatter, ) parser.add_argument("project_id", help="Google Cloud project ID") parser.add_argument("zone", help="GKE Cluster zone") diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt index cba90b27..05be978e 100644 --- a/samples/snippets/requirements.txt +++ b/samples/snippets/requirements.txt @@ -1,3 +1,3 @@ -google-cloud-container==2.10.6 -backoff==1.11.1 -pytest==7.1.1 \ No newline at end of file +google-cloud-container==2.10.7 +backoff==2.0.1 +pytest==7.1.2 \ No newline at end of file diff --git a/scripts/readme-gen/readme_gen.py b/scripts/readme-gen/readme_gen.py index d309d6e9..91b59676 100644 --- a/scripts/readme-gen/readme_gen.py +++ b/scripts/readme-gen/readme_gen.py @@ -28,7 +28,10 @@ jinja_env = jinja2.Environment( trim_blocks=True, loader=jinja2.FileSystemLoader( - os.path.abspath(os.path.join(os.path.dirname(__file__), 'templates')))) + os.path.abspath(os.path.join(os.path.dirname(__file__), "templates")) + ), + autoescape=True, +) README_TMPL = jinja_env.get_template('README.tmpl.rst') diff --git a/setup.py b/setup.py index 045f45aa..2ff2601c 100644 --- a/setup.py +++ b/setup.py @@ -17,12 +17,11 @@ import setuptools - # Package metadata. name = "google-cloud-container" description = "Google Container Engine API client library" -version = "2.10.7" +version = "2.10.8" # Should be one of: # 'Development Status :: 3 - Alpha' # 'Development Status :: 4 - Beta' @@ -33,8 +32,9 @@ # Until this issue is closed # https://github.com/googleapis/google-cloud-python/issues/10566 "google-api-core[grpc] >= 1.31.5, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.0", - "grpc-google-iam-v1 >= 0.12.3, < 0.13dev", - "proto-plus >= 1.15.0", + "grpc-google-iam-v1 >=0.12.4, <1.0.0dev", + "proto-plus >= 1.15.0, <2.0.0dev", + "protobuf >= 3.19.0, <4.0.0dev", ] extras = {"libcst": "libcst >= 0.2.5"} diff --git a/testing/constraints-3.6.txt b/testing/constraints-3.6.txt index 3755c967..c4e0fd7d 100644 --- a/testing/constraints-3.6.txt +++ b/testing/constraints-3.6.txt @@ -6,6 +6,7 @@ # e.g., if setup.py has "foo >= 1.14.0, < 2.0.0dev", # Then this file should have foo==1.14.0 google-api-core==1.31.5 -grpc-google-iam-v1==0.12.3 +grpc-google-iam-v1==0.12.4 proto-plus==1.15.0 libcst==0.2.5 +protobuf==3.19.0 diff --git a/testing/constraints-3.7.txt b/testing/constraints-3.7.txt index e69de29b..c4e0fd7d 100644 --- a/testing/constraints-3.7.txt +++ b/testing/constraints-3.7.txt @@ -0,0 +1,12 @@ +# This constraints file is used to check that lower bounds +# are correct in setup.py +# List *all* library dependencies and extras in this file. +# Pin the version to the lower bound. +# +# e.g., if setup.py has "foo >= 1.14.0, < 2.0.0dev", +# Then this file should have foo==1.14.0 +google-api-core==1.31.5 +grpc-google-iam-v1==0.12.4 +proto-plus==1.15.0 +libcst==0.2.5 +protobuf==3.19.0 diff --git a/tests/unit/gapic/container_v1/test_cluster_manager.py b/tests/unit/gapic/container_v1/test_cluster_manager.py index da1ed273..755bafd3 100644 --- a/tests/unit/gapic/container_v1/test_cluster_manager.py +++ b/tests/unit/gapic/container_v1/test_cluster_manager.py @@ -14,34 +14,39 @@ # limitations under the License. # import os -import mock -import grpc -from grpc.experimental import aio -import math -import pytest -from proto.marshal.rules.dates import DurationRule, TimestampRule +# try/except added for compatibility with python < 3.8 +try: + from unittest import mock + from unittest.mock import AsyncMock +except ImportError: + import mock +import math +from google.api_core import gapic_v1, grpc_helpers, grpc_helpers_async, path_template from google.api_core import client_options from google.api_core import exceptions as core_exceptions -from google.api_core import gapic_v1 -from google.api_core import grpc_helpers -from google.api_core import grpc_helpers_async -from google.api_core import path_template +import google.auth from google.auth import credentials as ga_credentials from google.auth.exceptions import MutualTLSChannelError -from google.cloud.container_v1.services.cluster_manager import ClusterManagerAsyncClient -from google.cloud.container_v1.services.cluster_manager import ClusterManagerClient -from google.cloud.container_v1.services.cluster_manager import pagers -from google.cloud.container_v1.services.cluster_manager import transports -from google.cloud.container_v1.types import cluster_service from google.oauth2 import service_account from google.protobuf import timestamp_pb2 # type: ignore from google.protobuf import wrappers_pb2 # type: ignore from google.rpc import code_pb2 # type: ignore from google.rpc import status_pb2 # type: ignore -import google.auth +import grpc +from grpc.experimental import aio +from proto.marshal.rules.dates import DurationRule, TimestampRule +import pytest + +from google.cloud.container_v1.services.cluster_manager import ( + ClusterManagerAsyncClient, + ClusterManagerClient, + pagers, + transports, +) +from google.cloud.container_v1.types import cluster_service def client_cert_source_callback(): @@ -89,20 +94,24 @@ def test__get_default_mtls_endpoint(): @pytest.mark.parametrize( - "client_class", [ClusterManagerClient, ClusterManagerAsyncClient,] + "client_class,transport_name", + [ + (ClusterManagerClient, "grpc"), + (ClusterManagerAsyncClient, "grpc_asyncio"), + ], ) -def test_cluster_manager_client_from_service_account_info(client_class): +def test_cluster_manager_client_from_service_account_info(client_class, transport_name): creds = ga_credentials.AnonymousCredentials() with mock.patch.object( service_account.Credentials, "from_service_account_info" ) as factory: factory.return_value = creds info = {"valid": True} - client = client_class.from_service_account_info(info) + client = client_class.from_service_account_info(info, transport=transport_name) assert client.transport._credentials == creds assert isinstance(client, client_class) - assert client.transport._host == "container.googleapis.com:443" + assert client.transport._host == ("container.googleapis.com:443") @pytest.mark.parametrize( @@ -131,23 +140,31 @@ def test_cluster_manager_client_service_account_always_use_jwt( @pytest.mark.parametrize( - "client_class", [ClusterManagerClient, ClusterManagerAsyncClient,] + "client_class,transport_name", + [ + (ClusterManagerClient, "grpc"), + (ClusterManagerAsyncClient, "grpc_asyncio"), + ], ) -def test_cluster_manager_client_from_service_account_file(client_class): +def test_cluster_manager_client_from_service_account_file(client_class, transport_name): creds = ga_credentials.AnonymousCredentials() with mock.patch.object( service_account.Credentials, "from_service_account_file" ) as factory: factory.return_value = creds - client = client_class.from_service_account_file("dummy/file/path.json") + client = client_class.from_service_account_file( + "dummy/file/path.json", transport=transport_name + ) assert client.transport._credentials == creds assert isinstance(client, client_class) - client = client_class.from_service_account_json("dummy/file/path.json") + client = client_class.from_service_account_json( + "dummy/file/path.json", transport=transport_name + ) assert client.transport._credentials == creds assert isinstance(client, client_class) - assert client.transport._host == "container.googleapis.com:443" + assert client.transport._host == ("container.googleapis.com:443") def test_cluster_manager_client_get_transport_class(): @@ -495,7 +512,9 @@ def test_cluster_manager_client_client_options_scopes( client_class, transport_class, transport_name ): # Check the case scopes are provided. - options = client_options.ClientOptions(scopes=["1", "2"],) + options = client_options.ClientOptions( + scopes=["1", "2"], + ) with mock.patch.object(transport_class, "__init__") as patched: patched.return_value = None client = client_class(client_options=options, transport=transport_name) @@ -635,10 +654,17 @@ def test_cluster_manager_client_create_channel_credentials_file( ) -@pytest.mark.parametrize("request_type", [cluster_service.ListClustersRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.ListClustersRequest, + dict, + ], +) def test_list_clusters(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -667,7 +693,8 @@ def test_list_clusters_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -683,7 +710,8 @@ async def test_list_clusters_async( transport: str = "grpc_asyncio", request_type=cluster_service.ListClustersRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -694,7 +722,9 @@ async def test_list_clusters_async( with mock.patch.object(type(client.transport.list_clusters), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( - cluster_service.ListClustersResponse(missing_zones=["missing_zones_value"],) + cluster_service.ListClustersResponse( + missing_zones=["missing_zones_value"], + ) ) response = await client.list_clusters(request) @@ -714,13 +744,15 @@ async def test_list_clusters_async_from_dict(): def test_list_clusters_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.ListClustersRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_clusters), "__call__") as call: @@ -734,7 +766,10 @@ def test_list_clusters_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -747,7 +782,7 @@ async def test_list_clusters_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.ListClustersRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_clusters), "__call__") as call: @@ -763,11 +798,16 @@ async def test_list_clusters_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_clusters_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_clusters), "__call__") as call: @@ -776,7 +816,9 @@ def test_list_clusters_flattened(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.list_clusters( - project_id="project_id_value", zone="zone_value", parent="parent_value", + project_id="project_id_value", + zone="zone_value", + parent="parent_value", ) # Establish that the underlying call was made with the expected @@ -795,7 +837,9 @@ def test_list_clusters_flattened(): def test_list_clusters_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -825,7 +869,9 @@ async def test_list_clusters_flattened_async(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.list_clusters( - project_id="project_id_value", zone="zone_value", parent="parent_value", + project_id="project_id_value", + zone="zone_value", + parent="parent_value", ) # Establish that the underlying call was made with the expected @@ -860,10 +906,17 @@ async def test_list_clusters_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.GetClusterRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.GetClusterRequest, + dict, + ], +) def test_get_cluster(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -948,7 +1001,8 @@ def test_get_cluster_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -964,7 +1018,8 @@ async def test_get_cluster_async( transport: str = "grpc_asyncio", request_type=cluster_service.GetClusterRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1053,13 +1108,15 @@ async def test_get_cluster_async_from_dict(): def test_get_cluster_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.GetClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_cluster), "__call__") as call: @@ -1073,7 +1130,10 @@ def test_get_cluster_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -1086,7 +1146,7 @@ async def test_get_cluster_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.GetClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_cluster), "__call__") as call: @@ -1102,11 +1162,16 @@ async def test_get_cluster_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_get_cluster_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_cluster), "__call__") as call: @@ -1140,7 +1205,9 @@ def test_get_cluster_flattened(): def test_get_cluster_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -1213,10 +1280,17 @@ async def test_get_cluster_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.CreateClusterRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.CreateClusterRequest, + dict, + ], +) def test_create_cluster(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1265,7 +1339,8 @@ def test_create_cluster_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -1281,7 +1356,8 @@ async def test_create_cluster_async( transport: str = "grpc_asyncio", request_type=cluster_service.CreateClusterRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1334,13 +1410,15 @@ async def test_create_cluster_async_from_dict(): def test_create_cluster_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.CreateClusterRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_cluster), "__call__") as call: @@ -1354,7 +1432,10 @@ def test_create_cluster_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -1367,7 +1448,7 @@ async def test_create_cluster_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.CreateClusterRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_cluster), "__call__") as call: @@ -1383,11 +1464,16 @@ async def test_create_cluster_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_create_cluster_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_cluster), "__call__") as call: @@ -1421,7 +1507,9 @@ def test_create_cluster_flattened(): def test_create_cluster_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -1494,10 +1582,17 @@ async def test_create_cluster_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.UpdateClusterRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.UpdateClusterRequest, + dict, + ], +) def test_update_cluster(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1546,7 +1641,8 @@ def test_update_cluster_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -1562,7 +1658,8 @@ async def test_update_cluster_async( transport: str = "grpc_asyncio", request_type=cluster_service.UpdateClusterRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1615,13 +1712,15 @@ async def test_update_cluster_async_from_dict(): def test_update_cluster_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.UpdateClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_cluster), "__call__") as call: @@ -1635,7 +1734,10 @@ def test_update_cluster_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -1648,7 +1750,7 @@ async def test_update_cluster_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.UpdateClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_cluster), "__call__") as call: @@ -1664,11 +1766,16 @@ async def test_update_cluster_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_update_cluster_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_cluster), "__call__") as call: @@ -1710,7 +1817,9 @@ def test_update_cluster_flattened(): def test_update_cluster_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -1797,10 +1906,17 @@ async def test_update_cluster_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.UpdateNodePoolRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.UpdateNodePoolRequest, + dict, + ], +) def test_update_node_pool(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1849,7 +1965,8 @@ def test_update_node_pool_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -1865,7 +1982,8 @@ async def test_update_node_pool_async( transport: str = "grpc_asyncio", request_type=cluster_service.UpdateNodePoolRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1918,13 +2036,15 @@ async def test_update_node_pool_async_from_dict(): def test_update_node_pool_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.UpdateNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_node_pool), "__call__") as call: @@ -1938,7 +2058,10 @@ def test_update_node_pool_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -1951,7 +2074,7 @@ async def test_update_node_pool_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.UpdateNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_node_pool), "__call__") as call: @@ -1967,15 +2090,23 @@ async def test_update_node_pool_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.parametrize( - "request_type", [cluster_service.SetNodePoolAutoscalingRequest, dict,] + "request_type", + [ + cluster_service.SetNodePoolAutoscalingRequest, + dict, + ], ) def test_set_node_pool_autoscaling(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2026,7 +2157,8 @@ def test_set_node_pool_autoscaling_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2045,7 +2177,8 @@ async def test_set_node_pool_autoscaling_async( request_type=cluster_service.SetNodePoolAutoscalingRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2100,13 +2233,15 @@ async def test_set_node_pool_autoscaling_async_from_dict(): def test_set_node_pool_autoscaling_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolAutoscalingRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2122,7 +2257,10 @@ def test_set_node_pool_autoscaling_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2135,7 +2273,7 @@ async def test_set_node_pool_autoscaling_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolAutoscalingRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2153,15 +2291,23 @@ async def test_set_node_pool_autoscaling_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.parametrize( - "request_type", [cluster_service.SetLoggingServiceRequest, dict,] + "request_type", + [ + cluster_service.SetLoggingServiceRequest, + dict, + ], ) def test_set_logging_service(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2212,7 +2358,8 @@ def test_set_logging_service_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2231,7 +2378,8 @@ async def test_set_logging_service_async( request_type=cluster_service.SetLoggingServiceRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2286,13 +2434,15 @@ async def test_set_logging_service_async_from_dict(): def test_set_logging_service_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetLoggingServiceRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2308,7 +2458,10 @@ def test_set_logging_service_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2321,7 +2474,7 @@ async def test_set_logging_service_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetLoggingServiceRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2339,11 +2492,16 @@ async def test_set_logging_service_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_logging_service_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2383,7 +2541,9 @@ def test_set_logging_service_flattened(): def test_set_logging_service_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -2465,11 +2625,16 @@ async def test_set_logging_service_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.SetMonitoringServiceRequest, dict,] + "request_type", + [ + cluster_service.SetMonitoringServiceRequest, + dict, + ], ) def test_set_monitoring_service(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2520,7 +2685,8 @@ def test_set_monitoring_service_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2539,7 +2705,8 @@ async def test_set_monitoring_service_async( request_type=cluster_service.SetMonitoringServiceRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2594,13 +2761,15 @@ async def test_set_monitoring_service_async_from_dict(): def test_set_monitoring_service_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetMonitoringServiceRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2616,7 +2785,10 @@ def test_set_monitoring_service_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2629,7 +2801,7 @@ async def test_set_monitoring_service_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetMonitoringServiceRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2647,11 +2819,16 @@ async def test_set_monitoring_service_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_monitoring_service_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2691,7 +2868,9 @@ def test_set_monitoring_service_flattened(): def test_set_monitoring_service_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -2773,11 +2952,16 @@ async def test_set_monitoring_service_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.SetAddonsConfigRequest, dict,] + "request_type", + [ + cluster_service.SetAddonsConfigRequest, + dict, + ], ) def test_set_addons_config(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2828,7 +3012,8 @@ def test_set_addons_config_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2846,7 +3031,8 @@ async def test_set_addons_config_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetAddonsConfigRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2901,13 +3087,15 @@ async def test_set_addons_config_async_from_dict(): def test_set_addons_config_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetAddonsConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2923,7 +3111,10 @@ def test_set_addons_config_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2936,7 +3127,7 @@ async def test_set_addons_config_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetAddonsConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2954,11 +3145,16 @@ async def test_set_addons_config_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_addons_config_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -3002,7 +3198,9 @@ def test_set_addons_config_flattened(): def test_set_addons_config_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -3091,10 +3289,17 @@ async def test_set_addons_config_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.SetLocationsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.SetLocationsRequest, + dict, + ], +) def test_set_locations(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3143,7 +3348,8 @@ def test_set_locations_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3159,7 +3365,8 @@ async def test_set_locations_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetLocationsRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3212,13 +3419,15 @@ async def test_set_locations_async_from_dict(): def test_set_locations_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetLocationsRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_locations), "__call__") as call: @@ -3232,7 +3441,10 @@ def test_set_locations_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -3245,7 +3457,7 @@ async def test_set_locations_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetLocationsRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_locations), "__call__") as call: @@ -3261,11 +3473,16 @@ async def test_set_locations_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_locations_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_locations), "__call__") as call: @@ -3303,7 +3520,9 @@ def test_set_locations_flattened(): def test_set_locations_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -3382,10 +3601,17 @@ async def test_set_locations_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.UpdateMasterRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.UpdateMasterRequest, + dict, + ], +) def test_update_master(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3434,7 +3660,8 @@ def test_update_master_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3450,7 +3677,8 @@ async def test_update_master_async( transport: str = "grpc_asyncio", request_type=cluster_service.UpdateMasterRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3503,13 +3731,15 @@ async def test_update_master_async_from_dict(): def test_update_master_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.UpdateMasterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_master), "__call__") as call: @@ -3523,7 +3753,10 @@ def test_update_master_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -3536,7 +3769,7 @@ async def test_update_master_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.UpdateMasterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_master), "__call__") as call: @@ -3552,11 +3785,16 @@ async def test_update_master_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_update_master_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_master), "__call__") as call: @@ -3594,7 +3832,9 @@ def test_update_master_flattened(): def test_update_master_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -3673,10 +3913,17 @@ async def test_update_master_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.SetMasterAuthRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.SetMasterAuthRequest, + dict, + ], +) def test_set_master_auth(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3725,7 +3972,8 @@ def test_set_master_auth_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3741,7 +3989,8 @@ async def test_set_master_auth_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetMasterAuthRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3794,13 +4043,15 @@ async def test_set_master_auth_async_from_dict(): def test_set_master_auth_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetMasterAuthRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_master_auth), "__call__") as call: @@ -3814,7 +4065,10 @@ def test_set_master_auth_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -3827,7 +4081,7 @@ async def test_set_master_auth_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetMasterAuthRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_master_auth), "__call__") as call: @@ -3843,13 +4097,23 @@ async def test_set_master_auth_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] -@pytest.mark.parametrize("request_type", [cluster_service.DeleteClusterRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.DeleteClusterRequest, + dict, + ], +) def test_delete_cluster(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3898,7 +4162,8 @@ def test_delete_cluster_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3914,7 +4179,8 @@ async def test_delete_cluster_async( transport: str = "grpc_asyncio", request_type=cluster_service.DeleteClusterRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3967,13 +4233,15 @@ async def test_delete_cluster_async_from_dict(): def test_delete_cluster_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.DeleteClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_cluster), "__call__") as call: @@ -3987,7 +4255,10 @@ def test_delete_cluster_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4000,7 +4271,7 @@ async def test_delete_cluster_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.DeleteClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_cluster), "__call__") as call: @@ -4016,11 +4287,16 @@ async def test_delete_cluster_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_delete_cluster_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_cluster), "__call__") as call: @@ -4054,7 +4330,9 @@ def test_delete_cluster_flattened(): def test_delete_cluster_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -4127,10 +4405,17 @@ async def test_delete_cluster_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.ListOperationsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.ListOperationsRequest, + dict, + ], +) def test_list_operations(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4159,7 +4444,8 @@ def test_list_operations_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4175,7 +4461,8 @@ async def test_list_operations_async( transport: str = "grpc_asyncio", request_type=cluster_service.ListOperationsRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4208,13 +4495,15 @@ async def test_list_operations_async_from_dict(): def test_list_operations_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.ListOperationsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_operations), "__call__") as call: @@ -4228,7 +4517,10 @@ def test_list_operations_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4241,7 +4533,7 @@ async def test_list_operations_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.ListOperationsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_operations), "__call__") as call: @@ -4257,11 +4549,16 @@ async def test_list_operations_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_operations_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_operations), "__call__") as call: @@ -4270,7 +4567,8 @@ def test_list_operations_flattened(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.list_operations( - project_id="project_id_value", zone="zone_value", + project_id="project_id_value", + zone="zone_value", ) # Establish that the underlying call was made with the expected @@ -4286,7 +4584,9 @@ def test_list_operations_flattened(): def test_list_operations_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -4315,7 +4615,8 @@ async def test_list_operations_flattened_async(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.list_operations( - project_id="project_id_value", zone="zone_value", + project_id="project_id_value", + zone="zone_value", ) # Establish that the underlying call was made with the expected @@ -4346,10 +4647,17 @@ async def test_list_operations_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.GetOperationRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.GetOperationRequest, + dict, + ], +) def test_get_operation(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4398,7 +4706,8 @@ def test_get_operation_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4414,7 +4723,8 @@ async def test_get_operation_async( transport: str = "grpc_asyncio", request_type=cluster_service.GetOperationRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4467,13 +4777,15 @@ async def test_get_operation_async_from_dict(): def test_get_operation_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.GetOperationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_operation), "__call__") as call: @@ -4487,7 +4799,10 @@ def test_get_operation_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4500,7 +4815,7 @@ async def test_get_operation_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.GetOperationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_operation), "__call__") as call: @@ -4516,11 +4831,16 @@ async def test_get_operation_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_get_operation_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_operation), "__call__") as call: @@ -4554,7 +4874,9 @@ def test_get_operation_flattened(): def test_get_operation_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -4628,11 +4950,16 @@ async def test_get_operation_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.CancelOperationRequest, dict,] + "request_type", + [ + cluster_service.CancelOperationRequest, + dict, + ], ) def test_cancel_operation(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4658,7 +4985,8 @@ def test_cancel_operation_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4674,7 +5002,8 @@ async def test_cancel_operation_async( transport: str = "grpc_asyncio", request_type=cluster_service.CancelOperationRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4702,13 +5031,15 @@ async def test_cancel_operation_async_from_dict(): def test_cancel_operation_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.CancelOperationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.cancel_operation), "__call__") as call: @@ -4722,7 +5053,10 @@ def test_cancel_operation_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4735,7 +5069,7 @@ async def test_cancel_operation_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.CancelOperationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.cancel_operation), "__call__") as call: @@ -4749,11 +5083,16 @@ async def test_cancel_operation_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_cancel_operation_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.cancel_operation), "__call__") as call: @@ -4787,7 +5126,9 @@ def test_cancel_operation_flattened(): def test_cancel_operation_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -4859,11 +5200,16 @@ async def test_cancel_operation_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.GetServerConfigRequest, dict,] + "request_type", + [ + cluster_service.GetServerConfigRequest, + dict, + ], ) def test_get_server_config(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4902,7 +5248,8 @@ def test_get_server_config_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4920,7 +5267,8 @@ async def test_get_server_config_async( transport: str = "grpc_asyncio", request_type=cluster_service.GetServerConfigRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4963,13 +5311,15 @@ async def test_get_server_config_async_from_dict(): def test_get_server_config_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.GetServerConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -4985,7 +5335,10 @@ def test_get_server_config_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4998,7 +5351,7 @@ async def test_get_server_config_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.GetServerConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -5016,11 +5369,16 @@ async def test_get_server_config_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_get_server_config_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -5031,7 +5389,9 @@ def test_get_server_config_flattened(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.get_server_config( - project_id="project_id_value", zone="zone_value", name="name_value", + project_id="project_id_value", + zone="zone_value", + name="name_value", ) # Establish that the underlying call was made with the expected @@ -5050,7 +5410,9 @@ def test_get_server_config_flattened(): def test_get_server_config_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -5082,7 +5444,9 @@ async def test_get_server_config_flattened_async(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.get_server_config( - project_id="project_id_value", zone="zone_value", name="name_value", + project_id="project_id_value", + zone="zone_value", + name="name_value", ) # Establish that the underlying call was made with the expected @@ -5117,10 +5481,17 @@ async def test_get_server_config_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.GetJSONWebKeysRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.GetJSONWebKeysRequest, + dict, + ], +) def test_get_json_web_keys(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5148,7 +5519,8 @@ def test_get_json_web_keys_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -5166,7 +5538,8 @@ async def test_get_json_web_keys_async( transport: str = "grpc_asyncio", request_type=cluster_service.GetJSONWebKeysRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5198,13 +5571,15 @@ async def test_get_json_web_keys_async_from_dict(): def test_get_json_web_keys_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.GetJSONWebKeysRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -5220,7 +5595,10 @@ def test_get_json_web_keys_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -5233,7 +5611,7 @@ async def test_get_json_web_keys_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.GetJSONWebKeysRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -5251,13 +5629,23 @@ async def test_get_json_web_keys_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] -@pytest.mark.parametrize("request_type", [cluster_service.ListNodePoolsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.ListNodePoolsRequest, + dict, + ], +) def test_list_node_pools(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5283,7 +5671,8 @@ def test_list_node_pools_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -5299,7 +5688,8 @@ async def test_list_node_pools_async( transport: str = "grpc_asyncio", request_type=cluster_service.ListNodePoolsRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5329,13 +5719,15 @@ async def test_list_node_pools_async_from_dict(): def test_list_node_pools_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.ListNodePoolsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_node_pools), "__call__") as call: @@ -5349,7 +5741,10 @@ def test_list_node_pools_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -5362,7 +5757,7 @@ async def test_list_node_pools_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.ListNodePoolsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_node_pools), "__call__") as call: @@ -5378,11 +5773,16 @@ async def test_list_node_pools_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_node_pools_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_node_pools), "__call__") as call: @@ -5416,7 +5816,9 @@ def test_list_node_pools_flattened(): def test_list_node_pools_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -5489,10 +5891,17 @@ async def test_list_node_pools_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.GetNodePoolRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.GetNodePoolRequest, + dict, + ], +) def test_get_node_pool(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5537,7 +5946,8 @@ def test_get_node_pool_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -5553,7 +5963,8 @@ async def test_get_node_pool_async( transport: str = "grpc_asyncio", request_type=cluster_service.GetNodePoolRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5602,13 +6013,15 @@ async def test_get_node_pool_async_from_dict(): def test_get_node_pool_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.GetNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_node_pool), "__call__") as call: @@ -5622,7 +6035,10 @@ def test_get_node_pool_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -5635,7 +6051,7 @@ async def test_get_node_pool_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.GetNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_node_pool), "__call__") as call: @@ -5651,11 +6067,16 @@ async def test_get_node_pool_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_get_node_pool_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_node_pool), "__call__") as call: @@ -5693,7 +6114,9 @@ def test_get_node_pool_flattened(): def test_get_node_pool_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -5772,10 +6195,17 @@ async def test_get_node_pool_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.CreateNodePoolRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.CreateNodePoolRequest, + dict, + ], +) def test_create_node_pool(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5824,7 +6254,8 @@ def test_create_node_pool_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -5840,7 +6271,8 @@ async def test_create_node_pool_async( transport: str = "grpc_asyncio", request_type=cluster_service.CreateNodePoolRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5893,13 +6325,15 @@ async def test_create_node_pool_async_from_dict(): def test_create_node_pool_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.CreateNodePoolRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_node_pool), "__call__") as call: @@ -5913,7 +6347,10 @@ def test_create_node_pool_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -5926,7 +6363,7 @@ async def test_create_node_pool_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.CreateNodePoolRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_node_pool), "__call__") as call: @@ -5942,11 +6379,16 @@ async def test_create_node_pool_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_create_node_pool_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_node_pool), "__call__") as call: @@ -5984,7 +6426,9 @@ def test_create_node_pool_flattened(): def test_create_node_pool_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -6063,10 +6507,17 @@ async def test_create_node_pool_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.DeleteNodePoolRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.DeleteNodePoolRequest, + dict, + ], +) def test_delete_node_pool(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6115,7 +6566,8 @@ def test_delete_node_pool_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -6131,7 +6583,8 @@ async def test_delete_node_pool_async( transport: str = "grpc_asyncio", request_type=cluster_service.DeleteNodePoolRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6184,13 +6637,15 @@ async def test_delete_node_pool_async_from_dict(): def test_delete_node_pool_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.DeleteNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_node_pool), "__call__") as call: @@ -6204,7 +6659,10 @@ def test_delete_node_pool_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -6217,7 +6675,7 @@ async def test_delete_node_pool_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.DeleteNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_node_pool), "__call__") as call: @@ -6233,11 +6691,16 @@ async def test_delete_node_pool_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_delete_node_pool_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_node_pool), "__call__") as call: @@ -6275,7 +6738,9 @@ def test_delete_node_pool_flattened(): def test_delete_node_pool_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -6355,11 +6820,16 @@ async def test_delete_node_pool_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.RollbackNodePoolUpgradeRequest, dict,] + "request_type", + [ + cluster_service.RollbackNodePoolUpgradeRequest, + dict, + ], ) def test_rollback_node_pool_upgrade(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6410,7 +6880,8 @@ def test_rollback_node_pool_upgrade_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -6429,7 +6900,8 @@ async def test_rollback_node_pool_upgrade_async( request_type=cluster_service.RollbackNodePoolUpgradeRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6484,13 +6956,15 @@ async def test_rollback_node_pool_upgrade_async_from_dict(): def test_rollback_node_pool_upgrade_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.RollbackNodePoolUpgradeRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -6506,7 +6980,10 @@ def test_rollback_node_pool_upgrade_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -6519,7 +6996,7 @@ async def test_rollback_node_pool_upgrade_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.RollbackNodePoolUpgradeRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -6537,11 +7014,16 @@ async def test_rollback_node_pool_upgrade_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_rollback_node_pool_upgrade_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -6581,7 +7063,9 @@ def test_rollback_node_pool_upgrade_flattened(): def test_rollback_node_pool_upgrade_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -6663,11 +7147,16 @@ async def test_rollback_node_pool_upgrade_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.SetNodePoolManagementRequest, dict,] + "request_type", + [ + cluster_service.SetNodePoolManagementRequest, + dict, + ], ) def test_set_node_pool_management(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6718,7 +7207,8 @@ def test_set_node_pool_management_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -6737,7 +7227,8 @@ async def test_set_node_pool_management_async( request_type=cluster_service.SetNodePoolManagementRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6792,13 +7283,15 @@ async def test_set_node_pool_management_async_from_dict(): def test_set_node_pool_management_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolManagementRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -6814,7 +7307,10 @@ def test_set_node_pool_management_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -6827,7 +7323,7 @@ async def test_set_node_pool_management_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolManagementRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -6845,13 +7341,23 @@ async def test_set_node_pool_management_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] -@pytest.mark.parametrize("request_type", [cluster_service.SetLabelsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.SetLabelsRequest, + dict, + ], +) def test_set_labels(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6900,7 +7406,8 @@ def test_set_labels_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -6916,7 +7423,8 @@ async def test_set_labels_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetLabelsRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6969,13 +7477,15 @@ async def test_set_labels_async_from_dict(): def test_set_labels_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetLabelsRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_labels), "__call__") as call: @@ -6989,7 +7499,10 @@ def test_set_labels_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -7002,7 +7515,7 @@ async def test_set_labels_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetLabelsRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_labels), "__call__") as call: @@ -7018,13 +7531,23 @@ async def test_set_labels_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] -@pytest.mark.parametrize("request_type", [cluster_service.SetLegacyAbacRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.SetLegacyAbacRequest, + dict, + ], +) def test_set_legacy_abac(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7073,7 +7596,8 @@ def test_set_legacy_abac_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -7089,7 +7613,8 @@ async def test_set_legacy_abac_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetLegacyAbacRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7142,13 +7667,15 @@ async def test_set_legacy_abac_async_from_dict(): def test_set_legacy_abac_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetLegacyAbacRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_legacy_abac), "__call__") as call: @@ -7162,7 +7689,10 @@ def test_set_legacy_abac_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -7175,7 +7705,7 @@ async def test_set_legacy_abac_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetLegacyAbacRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_legacy_abac), "__call__") as call: @@ -7191,11 +7721,16 @@ async def test_set_legacy_abac_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_legacy_abac_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_legacy_abac), "__call__") as call: @@ -7233,7 +7768,9 @@ def test_set_legacy_abac_flattened(): def test_set_legacy_abac_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -7313,11 +7850,16 @@ async def test_set_legacy_abac_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.StartIPRotationRequest, dict,] + "request_type", + [ + cluster_service.StartIPRotationRequest, + dict, + ], ) def test_start_ip_rotation(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7368,7 +7910,8 @@ def test_start_ip_rotation_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -7386,7 +7929,8 @@ async def test_start_ip_rotation_async( transport: str = "grpc_asyncio", request_type=cluster_service.StartIPRotationRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7441,13 +7985,15 @@ async def test_start_ip_rotation_async_from_dict(): def test_start_ip_rotation_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.StartIPRotationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7463,7 +8009,10 @@ def test_start_ip_rotation_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -7476,7 +8025,7 @@ async def test_start_ip_rotation_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.StartIPRotationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7494,11 +8043,16 @@ async def test_start_ip_rotation_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_start_ip_rotation_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7534,7 +8088,9 @@ def test_start_ip_rotation_flattened(): def test_start_ip_rotation_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -7610,11 +8166,16 @@ async def test_start_ip_rotation_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.CompleteIPRotationRequest, dict,] + "request_type", + [ + cluster_service.CompleteIPRotationRequest, + dict, + ], ) def test_complete_ip_rotation(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7665,7 +8226,8 @@ def test_complete_ip_rotation_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -7684,7 +8246,8 @@ async def test_complete_ip_rotation_async( request_type=cluster_service.CompleteIPRotationRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7739,13 +8302,15 @@ async def test_complete_ip_rotation_async_from_dict(): def test_complete_ip_rotation_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.CompleteIPRotationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7761,7 +8326,10 @@ def test_complete_ip_rotation_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -7774,7 +8342,7 @@ async def test_complete_ip_rotation_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.CompleteIPRotationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7792,11 +8360,16 @@ async def test_complete_ip_rotation_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_complete_ip_rotation_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7832,7 +8405,9 @@ def test_complete_ip_rotation_flattened(): def test_complete_ip_rotation_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -7908,11 +8483,16 @@ async def test_complete_ip_rotation_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.SetNodePoolSizeRequest, dict,] + "request_type", + [ + cluster_service.SetNodePoolSizeRequest, + dict, + ], ) def test_set_node_pool_size(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7963,7 +8543,8 @@ def test_set_node_pool_size_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -7981,7 +8562,8 @@ async def test_set_node_pool_size_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetNodePoolSizeRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8036,13 +8618,15 @@ async def test_set_node_pool_size_async_from_dict(): def test_set_node_pool_size_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolSizeRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8058,7 +8642,10 @@ def test_set_node_pool_size_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -8071,7 +8658,7 @@ async def test_set_node_pool_size_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolSizeRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8089,15 +8676,23 @@ async def test_set_node_pool_size_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.parametrize( - "request_type", [cluster_service.SetNetworkPolicyRequest, dict,] + "request_type", + [ + cluster_service.SetNetworkPolicyRequest, + dict, + ], ) def test_set_network_policy(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8148,7 +8743,8 @@ def test_set_network_policy_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -8167,7 +8763,8 @@ async def test_set_network_policy_async( request_type=cluster_service.SetNetworkPolicyRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8222,13 +8819,15 @@ async def test_set_network_policy_async_from_dict(): def test_set_network_policy_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetNetworkPolicyRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8244,7 +8843,10 @@ def test_set_network_policy_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -8257,7 +8859,7 @@ async def test_set_network_policy_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetNetworkPolicyRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8275,11 +8877,16 @@ async def test_set_network_policy_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_network_policy_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8323,7 +8930,9 @@ def test_set_network_policy_flattened(): def test_set_network_policy_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -8413,11 +9022,16 @@ async def test_set_network_policy_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.SetMaintenancePolicyRequest, dict,] + "request_type", + [ + cluster_service.SetMaintenancePolicyRequest, + dict, + ], ) def test_set_maintenance_policy(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8468,7 +9082,8 @@ def test_set_maintenance_policy_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -8487,7 +9102,8 @@ async def test_set_maintenance_policy_async( request_type=cluster_service.SetMaintenancePolicyRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8542,13 +9158,15 @@ async def test_set_maintenance_policy_async_from_dict(): def test_set_maintenance_policy_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetMaintenancePolicyRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8564,7 +9182,10 @@ def test_set_maintenance_policy_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -8577,7 +9198,7 @@ async def test_set_maintenance_policy_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetMaintenancePolicyRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8595,11 +9216,16 @@ async def test_set_maintenance_policy_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_maintenance_policy_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8651,7 +9277,9 @@ def test_set_maintenance_policy_flattened(): def test_set_maintenance_policy_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -8757,11 +9385,16 @@ async def test_set_maintenance_policy_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.ListUsableSubnetworksRequest, dict,] + "request_type", + [ + cluster_service.ListUsableSubnetworksRequest, + dict, + ], ) def test_list_usable_subnetworks(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8792,7 +9425,8 @@ def test_list_usable_subnetworks_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -8811,7 +9445,8 @@ async def test_list_usable_subnetworks_async( request_type=cluster_service.ListUsableSubnetworksRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8846,13 +9481,15 @@ async def test_list_usable_subnetworks_async_from_dict(): def test_list_usable_subnetworks_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.ListUsableSubnetworksRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8868,7 +9505,10 @@ def test_list_usable_subnetworks_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -8881,7 +9521,7 @@ async def test_list_usable_subnetworks_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.ListUsableSubnetworksRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8899,12 +9539,16 @@ async def test_list_usable_subnetworks_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_usable_subnetworks_pager(transport_name: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -8922,10 +9566,13 @@ def test_list_usable_subnetworks_pager(transport_name: str = "grpc"): next_page_token="abc", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[], next_page_token="def", + subnetworks=[], + next_page_token="def", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[cluster_service.UsableSubnetwork(),], + subnetworks=[ + cluster_service.UsableSubnetwork(), + ], next_page_token="ghi", ), cluster_service.ListUsableSubnetworksResponse( @@ -8945,14 +9592,15 @@ def test_list_usable_subnetworks_pager(transport_name: str = "grpc"): assert pager._metadata == metadata - results = [i for i in pager] + results = list(pager) assert len(results) == 6 assert all(isinstance(i, cluster_service.UsableSubnetwork) for i in results) def test_list_usable_subnetworks_pages(transport_name: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -8970,10 +9618,13 @@ def test_list_usable_subnetworks_pages(transport_name: str = "grpc"): next_page_token="abc", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[], next_page_token="def", + subnetworks=[], + next_page_token="def", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[cluster_service.UsableSubnetwork(),], + subnetworks=[ + cluster_service.UsableSubnetwork(), + ], next_page_token="ghi", ), cluster_service.ListUsableSubnetworksResponse( @@ -8991,7 +9642,9 @@ def test_list_usable_subnetworks_pages(transport_name: str = "grpc"): @pytest.mark.asyncio async def test_list_usable_subnetworks_async_pager(): - client = ClusterManagerAsyncClient(credentials=ga_credentials.AnonymousCredentials,) + client = ClusterManagerAsyncClient( + credentials=ga_credentials.AnonymousCredentials, + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -9010,10 +9663,13 @@ async def test_list_usable_subnetworks_async_pager(): next_page_token="abc", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[], next_page_token="def", + subnetworks=[], + next_page_token="def", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[cluster_service.UsableSubnetwork(),], + subnetworks=[ + cluster_service.UsableSubnetwork(), + ], next_page_token="ghi", ), cluster_service.ListUsableSubnetworksResponse( @@ -9024,10 +9680,12 @@ async def test_list_usable_subnetworks_async_pager(): ), RuntimeError, ) - async_pager = await client.list_usable_subnetworks(request={},) + async_pager = await client.list_usable_subnetworks( + request={}, + ) assert async_pager.next_page_token == "abc" responses = [] - async for response in async_pager: + async for response in async_pager: # pragma: no branch responses.append(response) assert len(responses) == 6 @@ -9036,7 +9694,9 @@ async def test_list_usable_subnetworks_async_pager(): @pytest.mark.asyncio async def test_list_usable_subnetworks_async_pages(): - client = ClusterManagerAsyncClient(credentials=ga_credentials.AnonymousCredentials,) + client = ClusterManagerAsyncClient( + credentials=ga_credentials.AnonymousCredentials, + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -9055,10 +9715,13 @@ async def test_list_usable_subnetworks_async_pages(): next_page_token="abc", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[], next_page_token="def", + subnetworks=[], + next_page_token="def", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[cluster_service.UsableSubnetwork(),], + subnetworks=[ + cluster_service.UsableSubnetwork(), + ], next_page_token="ghi", ), cluster_service.ListUsableSubnetworksResponse( @@ -9070,7 +9733,9 @@ async def test_list_usable_subnetworks_async_pages(): RuntimeError, ) pages = [] - async for page_ in (await client.list_usable_subnetworks(request={})).pages: + async for page_ in ( + await client.list_usable_subnetworks(request={}) + ).pages: # pragma: no branch pages.append(page_) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token @@ -9083,7 +9748,8 @@ def test_credentials_transport_error(): ) with pytest.raises(ValueError): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # It is an error to provide a credentials file and a transport instance. @@ -9103,7 +9769,10 @@ def test_credentials_transport_error(): options = client_options.ClientOptions() options.api_key = "api_key" with pytest.raises(ValueError): - client = ClusterManagerClient(client_options=options, transport=transport,) + client = ClusterManagerClient( + client_options=options, + transport=transport, + ) # It is an error to provide an api_key and a credential. options = mock.Mock() @@ -9119,7 +9788,8 @@ def test_credentials_transport_error(): ) with pytest.raises(ValueError): client = ClusterManagerClient( - client_options={"scopes": ["1", "2"]}, transport=transport, + client_options={"scopes": ["1", "2"]}, + transport=transport, ) @@ -9162,10 +9832,28 @@ def test_transport_adc(transport_class): adc.assert_called_once() +@pytest.mark.parametrize( + "transport_name", + [ + "grpc", + ], +) +def test_transport_kind(transport_name): + transport = ClusterManagerClient.get_transport_class(transport_name)( + credentials=ga_credentials.AnonymousCredentials(), + ) + assert transport.kind == transport_name + + def test_transport_grpc_default(): # A client should use the gRPC transport by default. - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) - assert isinstance(client.transport, transports.ClusterManagerGrpcTransport,) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) + assert isinstance( + client.transport, + transports.ClusterManagerGrpcTransport, + ) def test_cluster_manager_base_transport_error(): @@ -9230,6 +9918,14 @@ def test_cluster_manager_base_transport(): with pytest.raises(NotImplementedError): transport.close() + # Catch all for all remaining methods and properties + remainder = [ + "kind", + ] + for r in remainder: + with pytest.raises(NotImplementedError): + getattr(transport, r)() + def test_cluster_manager_base_transport_with_credentials_file(): # Instantiate the base transport with a credentials file @@ -9241,7 +9937,8 @@ def test_cluster_manager_base_transport_with_credentials_file(): Transport.return_value = None load_creds.return_value = (ga_credentials.AnonymousCredentials(), None) transport = transports.ClusterManagerTransport( - credentials_file="credentials.json", quota_project_id="octopus", + credentials_file="credentials.json", + quota_project_id="octopus", ) load_creds.assert_called_once_with( "credentials.json", @@ -9374,24 +10071,40 @@ def test_cluster_manager_grpc_transport_client_cert_source_for_mtls(transport_cl ) -def test_cluster_manager_host_no_port(): +@pytest.mark.parametrize( + "transport_name", + [ + "grpc", + "grpc_asyncio", + ], +) +def test_cluster_manager_host_no_port(transport_name): client = ClusterManagerClient( credentials=ga_credentials.AnonymousCredentials(), client_options=client_options.ClientOptions( api_endpoint="container.googleapis.com" ), + transport=transport_name, ) - assert client.transport._host == "container.googleapis.com:443" + assert client.transport._host == ("container.googleapis.com:443") -def test_cluster_manager_host_with_port(): +@pytest.mark.parametrize( + "transport_name", + [ + "grpc", + "grpc_asyncio", + ], +) +def test_cluster_manager_host_with_port(transport_name): client = ClusterManagerClient( credentials=ga_credentials.AnonymousCredentials(), client_options=client_options.ClientOptions( api_endpoint="container.googleapis.com:8000" ), + transport=transport_name, ) - assert client.transport._host == "container.googleapis.com:8000" + assert client.transport._host == ("container.googleapis.com:8000") def test_cluster_manager_grpc_transport_channel(): @@ -9399,7 +10112,8 @@ def test_cluster_manager_grpc_transport_channel(): # Check that channel is used if provided. transport = transports.ClusterManagerGrpcTransport( - host="squid.clam.whelk", channel=channel, + host="squid.clam.whelk", + channel=channel, ) assert transport.grpc_channel == channel assert transport._host == "squid.clam.whelk:443" @@ -9411,7 +10125,8 @@ def test_cluster_manager_grpc_asyncio_transport_channel(): # Check that channel is used if provided. transport = transports.ClusterManagerGrpcAsyncIOTransport( - host="squid.clam.whelk", channel=channel, + host="squid.clam.whelk", + channel=channel, ) assert transport.grpc_channel == channel assert transport._host == "squid.clam.whelk:443" @@ -9521,7 +10236,10 @@ def test_cluster_manager_transport_channel_mtls_with_adc(transport_class): def test_topic_path(): project = "squid" topic = "clam" - expected = "projects/{project}/topics/{topic}".format(project=project, topic=topic,) + expected = "projects/{project}/topics/{topic}".format( + project=project, + topic=topic, + ) actual = ClusterManagerClient.topic_path(project, topic) assert expected == actual @@ -9560,7 +10278,9 @@ def test_parse_common_billing_account_path(): def test_common_folder_path(): folder = "cuttlefish" - expected = "folders/{folder}".format(folder=folder,) + expected = "folders/{folder}".format( + folder=folder, + ) actual = ClusterManagerClient.common_folder_path(folder) assert expected == actual @@ -9578,7 +10298,9 @@ def test_parse_common_folder_path(): def test_common_organization_path(): organization = "winkle" - expected = "organizations/{organization}".format(organization=organization,) + expected = "organizations/{organization}".format( + organization=organization, + ) actual = ClusterManagerClient.common_organization_path(organization) assert expected == actual @@ -9596,7 +10318,9 @@ def test_parse_common_organization_path(): def test_common_project_path(): project = "scallop" - expected = "projects/{project}".format(project=project,) + expected = "projects/{project}".format( + project=project, + ) actual = ClusterManagerClient.common_project_path(project) assert expected == actual @@ -9616,7 +10340,8 @@ def test_common_location_path(): project = "squid" location = "clam" expected = "projects/{project}/locations/{location}".format( - project=project, location=location, + project=project, + location=location, ) actual = ClusterManagerClient.common_location_path(project, location) assert expected == actual @@ -9641,7 +10366,8 @@ def test_client_with_default_client_info(): transports.ClusterManagerTransport, "_prep_wrapped_messages" ) as prep: client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), client_info=client_info, + credentials=ga_credentials.AnonymousCredentials(), + client_info=client_info, ) prep.assert_called_once_with(client_info) @@ -9650,7 +10376,8 @@ def test_client_with_default_client_info(): ) as prep: transport_class = ClusterManagerClient.get_transport_class() transport = transport_class( - credentials=ga_credentials.AnonymousCredentials(), client_info=client_info, + credentials=ga_credentials.AnonymousCredentials(), + client_info=client_info, ) prep.assert_called_once_with(client_info) @@ -9658,7 +10385,8 @@ def test_client_with_default_client_info(): @pytest.mark.asyncio async def test_transport_close_async(): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc_asyncio", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc_asyncio", ) with mock.patch.object( type(getattr(client.transport, "grpc_channel")), "close" diff --git a/tests/unit/gapic/container_v1beta1/test_cluster_manager.py b/tests/unit/gapic/container_v1beta1/test_cluster_manager.py index f4dd2b4f..f06b1a11 100644 --- a/tests/unit/gapic/container_v1beta1/test_cluster_manager.py +++ b/tests/unit/gapic/container_v1beta1/test_cluster_manager.py @@ -14,36 +14,39 @@ # limitations under the License. # import os -import mock -import grpc -from grpc.experimental import aio -import math -import pytest -from proto.marshal.rules.dates import DurationRule, TimestampRule +# try/except added for compatibility with python < 3.8 +try: + from unittest import mock + from unittest.mock import AsyncMock +except ImportError: + import mock +import math +from google.api_core import gapic_v1, grpc_helpers, grpc_helpers_async, path_template from google.api_core import client_options from google.api_core import exceptions as core_exceptions -from google.api_core import gapic_v1 -from google.api_core import grpc_helpers -from google.api_core import grpc_helpers_async -from google.api_core import path_template +import google.auth from google.auth import credentials as ga_credentials from google.auth.exceptions import MutualTLSChannelError -from google.cloud.container_v1beta1.services.cluster_manager import ( - ClusterManagerAsyncClient, -) -from google.cloud.container_v1beta1.services.cluster_manager import ClusterManagerClient -from google.cloud.container_v1beta1.services.cluster_manager import pagers -from google.cloud.container_v1beta1.services.cluster_manager import transports -from google.cloud.container_v1beta1.types import cluster_service from google.oauth2 import service_account from google.protobuf import timestamp_pb2 # type: ignore from google.protobuf import wrappers_pb2 # type: ignore from google.rpc import code_pb2 # type: ignore from google.rpc import status_pb2 # type: ignore -import google.auth +import grpc +from grpc.experimental import aio +from proto.marshal.rules.dates import DurationRule, TimestampRule +import pytest + +from google.cloud.container_v1beta1.services.cluster_manager import ( + ClusterManagerAsyncClient, + ClusterManagerClient, + pagers, + transports, +) +from google.cloud.container_v1beta1.types import cluster_service def client_cert_source_callback(): @@ -91,20 +94,24 @@ def test__get_default_mtls_endpoint(): @pytest.mark.parametrize( - "client_class", [ClusterManagerClient, ClusterManagerAsyncClient,] + "client_class,transport_name", + [ + (ClusterManagerClient, "grpc"), + (ClusterManagerAsyncClient, "grpc_asyncio"), + ], ) -def test_cluster_manager_client_from_service_account_info(client_class): +def test_cluster_manager_client_from_service_account_info(client_class, transport_name): creds = ga_credentials.AnonymousCredentials() with mock.patch.object( service_account.Credentials, "from_service_account_info" ) as factory: factory.return_value = creds info = {"valid": True} - client = client_class.from_service_account_info(info) + client = client_class.from_service_account_info(info, transport=transport_name) assert client.transport._credentials == creds assert isinstance(client, client_class) - assert client.transport._host == "container.googleapis.com:443" + assert client.transport._host == ("container.googleapis.com:443") @pytest.mark.parametrize( @@ -133,23 +140,31 @@ def test_cluster_manager_client_service_account_always_use_jwt( @pytest.mark.parametrize( - "client_class", [ClusterManagerClient, ClusterManagerAsyncClient,] + "client_class,transport_name", + [ + (ClusterManagerClient, "grpc"), + (ClusterManagerAsyncClient, "grpc_asyncio"), + ], ) -def test_cluster_manager_client_from_service_account_file(client_class): +def test_cluster_manager_client_from_service_account_file(client_class, transport_name): creds = ga_credentials.AnonymousCredentials() with mock.patch.object( service_account.Credentials, "from_service_account_file" ) as factory: factory.return_value = creds - client = client_class.from_service_account_file("dummy/file/path.json") + client = client_class.from_service_account_file( + "dummy/file/path.json", transport=transport_name + ) assert client.transport._credentials == creds assert isinstance(client, client_class) - client = client_class.from_service_account_json("dummy/file/path.json") + client = client_class.from_service_account_json( + "dummy/file/path.json", transport=transport_name + ) assert client.transport._credentials == creds assert isinstance(client, client_class) - assert client.transport._host == "container.googleapis.com:443" + assert client.transport._host == ("container.googleapis.com:443") def test_cluster_manager_client_get_transport_class(): @@ -497,7 +512,9 @@ def test_cluster_manager_client_client_options_scopes( client_class, transport_class, transport_name ): # Check the case scopes are provided. - options = client_options.ClientOptions(scopes=["1", "2"],) + options = client_options.ClientOptions( + scopes=["1", "2"], + ) with mock.patch.object(transport_class, "__init__") as patched: patched.return_value = None client = client_class(client_options=options, transport=transport_name) @@ -637,10 +654,17 @@ def test_cluster_manager_client_create_channel_credentials_file( ) -@pytest.mark.parametrize("request_type", [cluster_service.ListClustersRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.ListClustersRequest, + dict, + ], +) def test_list_clusters(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -669,7 +693,8 @@ def test_list_clusters_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -685,7 +710,8 @@ async def test_list_clusters_async( transport: str = "grpc_asyncio", request_type=cluster_service.ListClustersRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -696,7 +722,9 @@ async def test_list_clusters_async( with mock.patch.object(type(client.transport.list_clusters), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( - cluster_service.ListClustersResponse(missing_zones=["missing_zones_value"],) + cluster_service.ListClustersResponse( + missing_zones=["missing_zones_value"], + ) ) response = await client.list_clusters(request) @@ -716,13 +744,15 @@ async def test_list_clusters_async_from_dict(): def test_list_clusters_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.ListClustersRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_clusters), "__call__") as call: @@ -736,7 +766,10 @@ def test_list_clusters_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -749,7 +782,7 @@ async def test_list_clusters_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.ListClustersRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_clusters), "__call__") as call: @@ -765,11 +798,16 @@ async def test_list_clusters_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_clusters_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_clusters), "__call__") as call: @@ -778,7 +816,8 @@ def test_list_clusters_flattened(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.list_clusters( - project_id="project_id_value", zone="zone_value", + project_id="project_id_value", + zone="zone_value", ) # Establish that the underlying call was made with the expected @@ -794,7 +833,9 @@ def test_list_clusters_flattened(): def test_list_clusters_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -823,7 +864,8 @@ async def test_list_clusters_flattened_async(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.list_clusters( - project_id="project_id_value", zone="zone_value", + project_id="project_id_value", + zone="zone_value", ) # Establish that the underlying call was made with the expected @@ -854,10 +896,17 @@ async def test_list_clusters_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.GetClusterRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.GetClusterRequest, + dict, + ], +) def test_get_cluster(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -946,7 +995,8 @@ def test_get_cluster_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -962,7 +1012,8 @@ async def test_get_cluster_async( transport: str = "grpc_asyncio", request_type=cluster_service.GetClusterRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1055,13 +1106,15 @@ async def test_get_cluster_async_from_dict(): def test_get_cluster_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.GetClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_cluster), "__call__") as call: @@ -1075,7 +1128,10 @@ def test_get_cluster_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -1088,7 +1144,7 @@ async def test_get_cluster_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.GetClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_cluster), "__call__") as call: @@ -1104,11 +1160,16 @@ async def test_get_cluster_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_get_cluster_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_cluster), "__call__") as call: @@ -1138,7 +1199,9 @@ def test_get_cluster_flattened(): def test_get_cluster_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -1205,10 +1268,17 @@ async def test_get_cluster_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.CreateClusterRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.CreateClusterRequest, + dict, + ], +) def test_create_cluster(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1257,7 +1327,8 @@ def test_create_cluster_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -1273,7 +1344,8 @@ async def test_create_cluster_async( transport: str = "grpc_asyncio", request_type=cluster_service.CreateClusterRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1326,13 +1398,15 @@ async def test_create_cluster_async_from_dict(): def test_create_cluster_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.CreateClusterRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_cluster), "__call__") as call: @@ -1346,7 +1420,10 @@ def test_create_cluster_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -1359,7 +1436,7 @@ async def test_create_cluster_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.CreateClusterRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_cluster), "__call__") as call: @@ -1375,11 +1452,16 @@ async def test_create_cluster_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_create_cluster_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_cluster), "__call__") as call: @@ -1409,7 +1491,9 @@ def test_create_cluster_flattened(): def test_create_cluster_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -1476,10 +1560,17 @@ async def test_create_cluster_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.UpdateClusterRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.UpdateClusterRequest, + dict, + ], +) def test_update_cluster(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1528,7 +1619,8 @@ def test_update_cluster_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -1544,7 +1636,8 @@ async def test_update_cluster_async( transport: str = "grpc_asyncio", request_type=cluster_service.UpdateClusterRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1597,13 +1690,15 @@ async def test_update_cluster_async_from_dict(): def test_update_cluster_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.UpdateClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_cluster), "__call__") as call: @@ -1617,7 +1712,10 @@ def test_update_cluster_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -1630,7 +1728,7 @@ async def test_update_cluster_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.UpdateClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_cluster), "__call__") as call: @@ -1646,11 +1744,16 @@ async def test_update_cluster_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_update_cluster_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_cluster), "__call__") as call: @@ -1688,7 +1791,9 @@ def test_update_cluster_flattened(): def test_update_cluster_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -1769,10 +1874,17 @@ async def test_update_cluster_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.UpdateNodePoolRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.UpdateNodePoolRequest, + dict, + ], +) def test_update_node_pool(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1821,7 +1933,8 @@ def test_update_node_pool_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -1837,7 +1950,8 @@ async def test_update_node_pool_async( transport: str = "grpc_asyncio", request_type=cluster_service.UpdateNodePoolRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1890,13 +2004,15 @@ async def test_update_node_pool_async_from_dict(): def test_update_node_pool_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.UpdateNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_node_pool), "__call__") as call: @@ -1910,7 +2026,10 @@ def test_update_node_pool_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -1923,7 +2042,7 @@ async def test_update_node_pool_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.UpdateNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_node_pool), "__call__") as call: @@ -1939,15 +2058,23 @@ async def test_update_node_pool_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.parametrize( - "request_type", [cluster_service.SetNodePoolAutoscalingRequest, dict,] + "request_type", + [ + cluster_service.SetNodePoolAutoscalingRequest, + dict, + ], ) def test_set_node_pool_autoscaling(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1998,7 +2125,8 @@ def test_set_node_pool_autoscaling_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2017,7 +2145,8 @@ async def test_set_node_pool_autoscaling_async( request_type=cluster_service.SetNodePoolAutoscalingRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2072,13 +2201,15 @@ async def test_set_node_pool_autoscaling_async_from_dict(): def test_set_node_pool_autoscaling_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolAutoscalingRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2094,7 +2225,10 @@ def test_set_node_pool_autoscaling_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2107,7 +2241,7 @@ async def test_set_node_pool_autoscaling_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolAutoscalingRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2125,15 +2259,23 @@ async def test_set_node_pool_autoscaling_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.parametrize( - "request_type", [cluster_service.SetLoggingServiceRequest, dict,] + "request_type", + [ + cluster_service.SetLoggingServiceRequest, + dict, + ], ) def test_set_logging_service(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2184,7 +2326,8 @@ def test_set_logging_service_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2203,7 +2346,8 @@ async def test_set_logging_service_async( request_type=cluster_service.SetLoggingServiceRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2258,13 +2402,15 @@ async def test_set_logging_service_async_from_dict(): def test_set_logging_service_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetLoggingServiceRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2280,7 +2426,10 @@ def test_set_logging_service_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2293,7 +2442,7 @@ async def test_set_logging_service_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetLoggingServiceRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2311,11 +2460,16 @@ async def test_set_logging_service_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_logging_service_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2351,7 +2505,9 @@ def test_set_logging_service_flattened(): def test_set_logging_service_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -2427,11 +2583,16 @@ async def test_set_logging_service_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.SetMonitoringServiceRequest, dict,] + "request_type", + [ + cluster_service.SetMonitoringServiceRequest, + dict, + ], ) def test_set_monitoring_service(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2482,7 +2643,8 @@ def test_set_monitoring_service_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2501,7 +2663,8 @@ async def test_set_monitoring_service_async( request_type=cluster_service.SetMonitoringServiceRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2556,13 +2719,15 @@ async def test_set_monitoring_service_async_from_dict(): def test_set_monitoring_service_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetMonitoringServiceRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2578,7 +2743,10 @@ def test_set_monitoring_service_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2591,7 +2759,7 @@ async def test_set_monitoring_service_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetMonitoringServiceRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2609,11 +2777,16 @@ async def test_set_monitoring_service_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_monitoring_service_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2649,7 +2822,9 @@ def test_set_monitoring_service_flattened(): def test_set_monitoring_service_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -2725,11 +2900,16 @@ async def test_set_monitoring_service_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.SetAddonsConfigRequest, dict,] + "request_type", + [ + cluster_service.SetAddonsConfigRequest, + dict, + ], ) def test_set_addons_config(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2780,7 +2960,8 @@ def test_set_addons_config_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2798,7 +2979,8 @@ async def test_set_addons_config_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetAddonsConfigRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2853,13 +3035,15 @@ async def test_set_addons_config_async_from_dict(): def test_set_addons_config_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetAddonsConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2875,7 +3059,10 @@ def test_set_addons_config_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2888,7 +3075,7 @@ async def test_set_addons_config_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetAddonsConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2906,11 +3093,16 @@ async def test_set_addons_config_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_addons_config_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2950,7 +3142,9 @@ def test_set_addons_config_flattened(): def test_set_addons_config_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -3033,10 +3227,17 @@ async def test_set_addons_config_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.SetLocationsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.SetLocationsRequest, + dict, + ], +) def test_set_locations(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3085,7 +3286,8 @@ def test_set_locations_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3101,7 +3303,8 @@ async def test_set_locations_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetLocationsRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3154,13 +3357,15 @@ async def test_set_locations_async_from_dict(): def test_set_locations_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetLocationsRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_locations), "__call__") as call: @@ -3174,7 +3379,10 @@ def test_set_locations_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -3187,7 +3395,7 @@ async def test_set_locations_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetLocationsRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_locations), "__call__") as call: @@ -3203,11 +3411,16 @@ async def test_set_locations_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_locations_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_locations), "__call__") as call: @@ -3241,7 +3454,9 @@ def test_set_locations_flattened(): def test_set_locations_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -3314,10 +3529,17 @@ async def test_set_locations_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.UpdateMasterRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.UpdateMasterRequest, + dict, + ], +) def test_update_master(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3366,7 +3588,8 @@ def test_update_master_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3382,7 +3605,8 @@ async def test_update_master_async( transport: str = "grpc_asyncio", request_type=cluster_service.UpdateMasterRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3435,13 +3659,15 @@ async def test_update_master_async_from_dict(): def test_update_master_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.UpdateMasterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_master), "__call__") as call: @@ -3455,7 +3681,10 @@ def test_update_master_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -3468,7 +3697,7 @@ async def test_update_master_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.UpdateMasterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_master), "__call__") as call: @@ -3484,11 +3713,16 @@ async def test_update_master_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_update_master_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_master), "__call__") as call: @@ -3522,7 +3756,9 @@ def test_update_master_flattened(): def test_update_master_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -3595,10 +3831,17 @@ async def test_update_master_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.SetMasterAuthRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.SetMasterAuthRequest, + dict, + ], +) def test_set_master_auth(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3647,7 +3890,8 @@ def test_set_master_auth_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3663,7 +3907,8 @@ async def test_set_master_auth_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetMasterAuthRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3716,13 +3961,15 @@ async def test_set_master_auth_async_from_dict(): def test_set_master_auth_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetMasterAuthRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_master_auth), "__call__") as call: @@ -3736,7 +3983,10 @@ def test_set_master_auth_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -3749,7 +3999,7 @@ async def test_set_master_auth_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetMasterAuthRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_master_auth), "__call__") as call: @@ -3765,13 +4015,23 @@ async def test_set_master_auth_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] -@pytest.mark.parametrize("request_type", [cluster_service.DeleteClusterRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.DeleteClusterRequest, + dict, + ], +) def test_delete_cluster(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3820,7 +4080,8 @@ def test_delete_cluster_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3836,7 +4097,8 @@ async def test_delete_cluster_async( transport: str = "grpc_asyncio", request_type=cluster_service.DeleteClusterRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3889,13 +4151,15 @@ async def test_delete_cluster_async_from_dict(): def test_delete_cluster_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.DeleteClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_cluster), "__call__") as call: @@ -3909,7 +4173,10 @@ def test_delete_cluster_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -3922,7 +4189,7 @@ async def test_delete_cluster_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.DeleteClusterRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_cluster), "__call__") as call: @@ -3938,11 +4205,16 @@ async def test_delete_cluster_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_delete_cluster_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_cluster), "__call__") as call: @@ -3972,7 +4244,9 @@ def test_delete_cluster_flattened(): def test_delete_cluster_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -4039,10 +4313,17 @@ async def test_delete_cluster_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.ListOperationsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.ListOperationsRequest, + dict, + ], +) def test_list_operations(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4071,7 +4352,8 @@ def test_list_operations_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4087,7 +4369,8 @@ async def test_list_operations_async( transport: str = "grpc_asyncio", request_type=cluster_service.ListOperationsRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4120,13 +4403,15 @@ async def test_list_operations_async_from_dict(): def test_list_operations_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.ListOperationsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_operations), "__call__") as call: @@ -4140,7 +4425,10 @@ def test_list_operations_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4153,7 +4441,7 @@ async def test_list_operations_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.ListOperationsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_operations), "__call__") as call: @@ -4169,12 +4457,17 @@ async def test_list_operations_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_operations_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) - + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) + # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_operations), "__call__") as call: # Designate an appropriate return value for the call. @@ -4182,7 +4475,8 @@ def test_list_operations_flattened(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.list_operations( - project_id="project_id_value", zone="zone_value", + project_id="project_id_value", + zone="zone_value", ) # Establish that the underlying call was made with the expected @@ -4198,7 +4492,9 @@ def test_list_operations_flattened(): def test_list_operations_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -4227,7 +4523,8 @@ async def test_list_operations_flattened_async(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.list_operations( - project_id="project_id_value", zone="zone_value", + project_id="project_id_value", + zone="zone_value", ) # Establish that the underlying call was made with the expected @@ -4258,10 +4555,17 @@ async def test_list_operations_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.GetOperationRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.GetOperationRequest, + dict, + ], +) def test_get_operation(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4310,7 +4614,8 @@ def test_get_operation_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4326,7 +4631,8 @@ async def test_get_operation_async( transport: str = "grpc_asyncio", request_type=cluster_service.GetOperationRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4379,13 +4685,15 @@ async def test_get_operation_async_from_dict(): def test_get_operation_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.GetOperationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_operation), "__call__") as call: @@ -4399,7 +4707,10 @@ def test_get_operation_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4412,7 +4723,7 @@ async def test_get_operation_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.GetOperationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_operation), "__call__") as call: @@ -4428,11 +4739,16 @@ async def test_get_operation_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_get_operation_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_operation), "__call__") as call: @@ -4462,7 +4778,9 @@ def test_get_operation_flattened(): def test_get_operation_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -4530,11 +4848,16 @@ async def test_get_operation_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.CancelOperationRequest, dict,] + "request_type", + [ + cluster_service.CancelOperationRequest, + dict, + ], ) def test_cancel_operation(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4560,7 +4883,8 @@ def test_cancel_operation_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4576,7 +4900,8 @@ async def test_cancel_operation_async( transport: str = "grpc_asyncio", request_type=cluster_service.CancelOperationRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4604,13 +4929,15 @@ async def test_cancel_operation_async_from_dict(): def test_cancel_operation_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.CancelOperationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.cancel_operation), "__call__") as call: @@ -4624,7 +4951,10 @@ def test_cancel_operation_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4637,7 +4967,7 @@ async def test_cancel_operation_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.CancelOperationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.cancel_operation), "__call__") as call: @@ -4651,11 +4981,16 @@ async def test_cancel_operation_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_cancel_operation_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.cancel_operation), "__call__") as call: @@ -4685,7 +5020,9 @@ def test_cancel_operation_flattened(): def test_cancel_operation_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -4751,11 +5088,16 @@ async def test_cancel_operation_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.GetServerConfigRequest, dict,] + "request_type", + [ + cluster_service.GetServerConfigRequest, + dict, + ], ) def test_get_server_config(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4794,7 +5136,8 @@ def test_get_server_config_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4812,7 +5155,8 @@ async def test_get_server_config_async( transport: str = "grpc_asyncio", request_type=cluster_service.GetServerConfigRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4855,13 +5199,15 @@ async def test_get_server_config_async_from_dict(): def test_get_server_config_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.GetServerConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -4877,7 +5223,10 @@ def test_get_server_config_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4890,7 +5239,7 @@ async def test_get_server_config_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.GetServerConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -4908,11 +5257,16 @@ async def test_get_server_config_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_get_server_config_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -4923,7 +5277,8 @@ def test_get_server_config_flattened(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.get_server_config( - project_id="project_id_value", zone="zone_value", + project_id="project_id_value", + zone="zone_value", ) # Establish that the underlying call was made with the expected @@ -4939,7 +5294,9 @@ def test_get_server_config_flattened(): def test_get_server_config_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -4970,7 +5327,8 @@ async def test_get_server_config_flattened_async(): # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.get_server_config( - project_id="project_id_value", zone="zone_value", + project_id="project_id_value", + zone="zone_value", ) # Establish that the underlying call was made with the expected @@ -5001,10 +5359,17 @@ async def test_get_server_config_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.ListNodePoolsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.ListNodePoolsRequest, + dict, + ], +) def test_list_node_pools(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5030,7 +5395,8 @@ def test_list_node_pools_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -5046,7 +5412,8 @@ async def test_list_node_pools_async( transport: str = "grpc_asyncio", request_type=cluster_service.ListNodePoolsRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5076,13 +5443,15 @@ async def test_list_node_pools_async_from_dict(): def test_list_node_pools_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.ListNodePoolsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_node_pools), "__call__") as call: @@ -5096,7 +5465,10 @@ def test_list_node_pools_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -5109,7 +5481,7 @@ async def test_list_node_pools_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.ListNodePoolsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_node_pools), "__call__") as call: @@ -5125,11 +5497,16 @@ async def test_list_node_pools_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_node_pools_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_node_pools), "__call__") as call: @@ -5159,7 +5536,9 @@ def test_list_node_pools_flattened(): def test_list_node_pools_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -5226,10 +5605,17 @@ async def test_list_node_pools_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.GetJSONWebKeysRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.GetJSONWebKeysRequest, + dict, + ], +) def test_get_json_web_keys(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5257,7 +5643,8 @@ def test_get_json_web_keys_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -5275,7 +5662,8 @@ async def test_get_json_web_keys_async( transport: str = "grpc_asyncio", request_type=cluster_service.GetJSONWebKeysRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5307,13 +5695,15 @@ async def test_get_json_web_keys_async_from_dict(): def test_get_json_web_keys_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.GetJSONWebKeysRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -5329,7 +5719,10 @@ def test_get_json_web_keys_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -5342,7 +5735,7 @@ async def test_get_json_web_keys_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.GetJSONWebKeysRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -5360,13 +5753,23 @@ async def test_get_json_web_keys_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] -@pytest.mark.parametrize("request_type", [cluster_service.GetNodePoolRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.GetNodePoolRequest, + dict, + ], +) def test_get_node_pool(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5411,7 +5814,8 @@ def test_get_node_pool_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -5427,7 +5831,8 @@ async def test_get_node_pool_async( transport: str = "grpc_asyncio", request_type=cluster_service.GetNodePoolRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5476,13 +5881,15 @@ async def test_get_node_pool_async_from_dict(): def test_get_node_pool_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.GetNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_node_pool), "__call__") as call: @@ -5496,7 +5903,10 @@ def test_get_node_pool_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -5509,7 +5919,7 @@ async def test_get_node_pool_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.GetNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_node_pool), "__call__") as call: @@ -5525,11 +5935,16 @@ async def test_get_node_pool_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_get_node_pool_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_node_pool), "__call__") as call: @@ -5563,7 +5978,9 @@ def test_get_node_pool_flattened(): def test_get_node_pool_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -5636,10 +6053,17 @@ async def test_get_node_pool_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.CreateNodePoolRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.CreateNodePoolRequest, + dict, + ], +) def test_create_node_pool(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5688,7 +6112,8 @@ def test_create_node_pool_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -5704,7 +6129,8 @@ async def test_create_node_pool_async( transport: str = "grpc_asyncio", request_type=cluster_service.CreateNodePoolRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5757,13 +6183,15 @@ async def test_create_node_pool_async_from_dict(): def test_create_node_pool_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.CreateNodePoolRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_node_pool), "__call__") as call: @@ -5777,7 +6205,10 @@ def test_create_node_pool_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -5790,7 +6221,7 @@ async def test_create_node_pool_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.CreateNodePoolRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_node_pool), "__call__") as call: @@ -5806,11 +6237,16 @@ async def test_create_node_pool_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_create_node_pool_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_node_pool), "__call__") as call: @@ -5844,7 +6280,9 @@ def test_create_node_pool_flattened(): def test_create_node_pool_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -5917,10 +6355,17 @@ async def test_create_node_pool_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.DeleteNodePoolRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.DeleteNodePoolRequest, + dict, + ], +) def test_delete_node_pool(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -5969,7 +6414,8 @@ def test_delete_node_pool_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -5985,7 +6431,8 @@ async def test_delete_node_pool_async( transport: str = "grpc_asyncio", request_type=cluster_service.DeleteNodePoolRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6038,13 +6485,15 @@ async def test_delete_node_pool_async_from_dict(): def test_delete_node_pool_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.DeleteNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_node_pool), "__call__") as call: @@ -6058,7 +6507,10 @@ def test_delete_node_pool_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -6071,7 +6523,7 @@ async def test_delete_node_pool_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.DeleteNodePoolRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_node_pool), "__call__") as call: @@ -6087,11 +6539,16 @@ async def test_delete_node_pool_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_delete_node_pool_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_node_pool), "__call__") as call: @@ -6125,7 +6582,9 @@ def test_delete_node_pool_flattened(): def test_delete_node_pool_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -6199,11 +6658,16 @@ async def test_delete_node_pool_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.RollbackNodePoolUpgradeRequest, dict,] + "request_type", + [ + cluster_service.RollbackNodePoolUpgradeRequest, + dict, + ], ) def test_rollback_node_pool_upgrade(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6254,7 +6718,8 @@ def test_rollback_node_pool_upgrade_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -6273,7 +6738,8 @@ async def test_rollback_node_pool_upgrade_async( request_type=cluster_service.RollbackNodePoolUpgradeRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6328,13 +6794,15 @@ async def test_rollback_node_pool_upgrade_async_from_dict(): def test_rollback_node_pool_upgrade_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.RollbackNodePoolUpgradeRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -6350,7 +6818,10 @@ def test_rollback_node_pool_upgrade_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -6363,7 +6834,7 @@ async def test_rollback_node_pool_upgrade_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.RollbackNodePoolUpgradeRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -6381,11 +6852,16 @@ async def test_rollback_node_pool_upgrade_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_rollback_node_pool_upgrade_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -6421,7 +6897,9 @@ def test_rollback_node_pool_upgrade_flattened(): def test_rollback_node_pool_upgrade_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -6497,11 +6975,16 @@ async def test_rollback_node_pool_upgrade_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.SetNodePoolManagementRequest, dict,] + "request_type", + [ + cluster_service.SetNodePoolManagementRequest, + dict, + ], ) def test_set_node_pool_management(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6552,7 +7035,8 @@ def test_set_node_pool_management_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -6571,7 +7055,8 @@ async def test_set_node_pool_management_async( request_type=cluster_service.SetNodePoolManagementRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6626,13 +7111,15 @@ async def test_set_node_pool_management_async_from_dict(): def test_set_node_pool_management_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolManagementRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -6648,7 +7135,10 @@ def test_set_node_pool_management_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -6661,7 +7151,7 @@ async def test_set_node_pool_management_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolManagementRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -6679,11 +7169,16 @@ async def test_set_node_pool_management_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_node_pool_management_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -6723,7 +7218,9 @@ def test_set_node_pool_management_flattened(): def test_set_node_pool_management_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -6804,10 +7301,17 @@ async def test_set_node_pool_management_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.SetLabelsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.SetLabelsRequest, + dict, + ], +) def test_set_labels(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6856,7 +7360,8 @@ def test_set_labels_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -6872,7 +7377,8 @@ async def test_set_labels_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetLabelsRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -6925,13 +7431,15 @@ async def test_set_labels_async_from_dict(): def test_set_labels_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetLabelsRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_labels), "__call__") as call: @@ -6945,7 +7453,10 @@ def test_set_labels_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -6958,7 +7469,7 @@ async def test_set_labels_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetLabelsRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_labels), "__call__") as call: @@ -6974,11 +7485,16 @@ async def test_set_labels_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_labels_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_labels), "__call__") as call: @@ -7016,7 +7532,9 @@ def test_set_labels_flattened(): def test_set_labels_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -7095,10 +7613,17 @@ async def test_set_labels_flattened_error_async(): ) -@pytest.mark.parametrize("request_type", [cluster_service.SetLegacyAbacRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.SetLegacyAbacRequest, + dict, + ], +) def test_set_legacy_abac(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7147,7 +7672,8 @@ def test_set_legacy_abac_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -7163,7 +7689,8 @@ async def test_set_legacy_abac_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetLegacyAbacRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7216,13 +7743,15 @@ async def test_set_legacy_abac_async_from_dict(): def test_set_legacy_abac_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetLegacyAbacRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_legacy_abac), "__call__") as call: @@ -7236,7 +7765,10 @@ def test_set_legacy_abac_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -7249,7 +7781,7 @@ async def test_set_legacy_abac_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetLegacyAbacRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_legacy_abac), "__call__") as call: @@ -7265,11 +7797,16 @@ async def test_set_legacy_abac_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_legacy_abac_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.set_legacy_abac), "__call__") as call: @@ -7303,7 +7840,9 @@ def test_set_legacy_abac_flattened(): def test_set_legacy_abac_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -7377,11 +7916,16 @@ async def test_set_legacy_abac_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.StartIPRotationRequest, dict,] + "request_type", + [ + cluster_service.StartIPRotationRequest, + dict, + ], ) def test_start_ip_rotation(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7432,7 +7976,8 @@ def test_start_ip_rotation_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -7450,7 +7995,8 @@ async def test_start_ip_rotation_async( transport: str = "grpc_asyncio", request_type=cluster_service.StartIPRotationRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7505,13 +8051,15 @@ async def test_start_ip_rotation_async_from_dict(): def test_start_ip_rotation_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.StartIPRotationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7527,7 +8075,10 @@ def test_start_ip_rotation_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -7540,7 +8091,7 @@ async def test_start_ip_rotation_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.StartIPRotationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7558,11 +8109,16 @@ async def test_start_ip_rotation_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_start_ip_rotation_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7594,7 +8150,9 @@ def test_start_ip_rotation_flattened(): def test_start_ip_rotation_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -7664,11 +8222,16 @@ async def test_start_ip_rotation_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.CompleteIPRotationRequest, dict,] + "request_type", + [ + cluster_service.CompleteIPRotationRequest, + dict, + ], ) def test_complete_ip_rotation(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7719,7 +8282,8 @@ def test_complete_ip_rotation_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -7738,7 +8302,8 @@ async def test_complete_ip_rotation_async( request_type=cluster_service.CompleteIPRotationRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -7793,13 +8358,15 @@ async def test_complete_ip_rotation_async_from_dict(): def test_complete_ip_rotation_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.CompleteIPRotationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7815,7 +8382,10 @@ def test_complete_ip_rotation_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -7828,7 +8398,7 @@ async def test_complete_ip_rotation_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.CompleteIPRotationRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7846,11 +8416,16 @@ async def test_complete_ip_rotation_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_complete_ip_rotation_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -7882,7 +8457,9 @@ def test_complete_ip_rotation_flattened(): def test_complete_ip_rotation_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -7952,11 +8529,16 @@ async def test_complete_ip_rotation_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.SetNodePoolSizeRequest, dict,] + "request_type", + [ + cluster_service.SetNodePoolSizeRequest, + dict, + ], ) def test_set_node_pool_size(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8007,7 +8589,8 @@ def test_set_node_pool_size_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -8025,7 +8608,8 @@ async def test_set_node_pool_size_async( transport: str = "grpc_asyncio", request_type=cluster_service.SetNodePoolSizeRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8080,13 +8664,15 @@ async def test_set_node_pool_size_async_from_dict(): def test_set_node_pool_size_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolSizeRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8102,7 +8688,10 @@ def test_set_node_pool_size_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -8115,7 +8704,7 @@ async def test_set_node_pool_size_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetNodePoolSizeRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8133,15 +8722,23 @@ async def test_set_node_pool_size_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.parametrize( - "request_type", [cluster_service.SetNetworkPolicyRequest, dict,] + "request_type", + [ + cluster_service.SetNetworkPolicyRequest, + dict, + ], ) def test_set_network_policy(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8192,7 +8789,8 @@ def test_set_network_policy_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -8211,7 +8809,8 @@ async def test_set_network_policy_async( request_type=cluster_service.SetNetworkPolicyRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8266,13 +8865,15 @@ async def test_set_network_policy_async_from_dict(): def test_set_network_policy_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetNetworkPolicyRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8288,7 +8889,10 @@ def test_set_network_policy_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -8301,7 +8905,7 @@ async def test_set_network_policy_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetNetworkPolicyRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8319,11 +8923,16 @@ async def test_set_network_policy_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_network_policy_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8363,7 +8972,9 @@ def test_set_network_policy_flattened(): def test_set_network_policy_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -8447,11 +9058,16 @@ async def test_set_network_policy_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.SetMaintenancePolicyRequest, dict,] + "request_type", + [ + cluster_service.SetMaintenancePolicyRequest, + dict, + ], ) def test_set_maintenance_policy(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8502,7 +9118,8 @@ def test_set_maintenance_policy_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -8521,7 +9138,8 @@ async def test_set_maintenance_policy_async( request_type=cluster_service.SetMaintenancePolicyRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8576,13 +9194,15 @@ async def test_set_maintenance_policy_async_from_dict(): def test_set_maintenance_policy_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.SetMaintenancePolicyRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8598,7 +9218,10 @@ def test_set_maintenance_policy_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -8611,7 +9234,7 @@ async def test_set_maintenance_policy_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.SetMaintenancePolicyRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8629,11 +9252,16 @@ async def test_set_maintenance_policy_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_set_maintenance_policy_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8681,7 +9309,9 @@ def test_set_maintenance_policy_flattened(): def test_set_maintenance_policy_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. @@ -8781,11 +9411,16 @@ async def test_set_maintenance_policy_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [cluster_service.ListUsableSubnetworksRequest, dict,] + "request_type", + [ + cluster_service.ListUsableSubnetworksRequest, + dict, + ], ) def test_list_usable_subnetworks(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8816,7 +9451,8 @@ def test_list_usable_subnetworks_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -8835,7 +9471,8 @@ async def test_list_usable_subnetworks_async( request_type=cluster_service.ListUsableSubnetworksRequest, ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -8870,13 +9507,15 @@ async def test_list_usable_subnetworks_async_from_dict(): def test_list_usable_subnetworks_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.ListUsableSubnetworksRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8892,7 +9531,10 @@ def test_list_usable_subnetworks_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -8905,7 +9547,7 @@ async def test_list_usable_subnetworks_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.ListUsableSubnetworksRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8923,11 +9565,16 @@ async def test_list_usable_subnetworks_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_usable_subnetworks_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -8937,7 +9584,9 @@ def test_list_usable_subnetworks_flattened(): call.return_value = cluster_service.ListUsableSubnetworksResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.list_usable_subnetworks(parent="parent_value",) + client.list_usable_subnetworks( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -8949,13 +9598,16 @@ def test_list_usable_subnetworks_flattened(): def test_list_usable_subnetworks_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.list_usable_subnetworks( - cluster_service.ListUsableSubnetworksRequest(), parent="parent_value", + cluster_service.ListUsableSubnetworksRequest(), + parent="parent_value", ) @@ -8977,7 +9629,9 @@ async def test_list_usable_subnetworks_flattened_async(): ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.list_usable_subnetworks(parent="parent_value",) + response = await client.list_usable_subnetworks( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -8998,13 +9652,15 @@ async def test_list_usable_subnetworks_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.list_usable_subnetworks( - cluster_service.ListUsableSubnetworksRequest(), parent="parent_value", + cluster_service.ListUsableSubnetworksRequest(), + parent="parent_value", ) def test_list_usable_subnetworks_pager(transport_name: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -9022,10 +9678,13 @@ def test_list_usable_subnetworks_pager(transport_name: str = "grpc"): next_page_token="abc", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[], next_page_token="def", + subnetworks=[], + next_page_token="def", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[cluster_service.UsableSubnetwork(),], + subnetworks=[ + cluster_service.UsableSubnetwork(), + ], next_page_token="ghi", ), cluster_service.ListUsableSubnetworksResponse( @@ -9045,14 +9704,15 @@ def test_list_usable_subnetworks_pager(transport_name: str = "grpc"): assert pager._metadata == metadata - results = [i for i in pager] + results = list(pager) assert len(results) == 6 assert all(isinstance(i, cluster_service.UsableSubnetwork) for i in results) def test_list_usable_subnetworks_pages(transport_name: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -9070,10 +9730,13 @@ def test_list_usable_subnetworks_pages(transport_name: str = "grpc"): next_page_token="abc", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[], next_page_token="def", + subnetworks=[], + next_page_token="def", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[cluster_service.UsableSubnetwork(),], + subnetworks=[ + cluster_service.UsableSubnetwork(), + ], next_page_token="ghi", ), cluster_service.ListUsableSubnetworksResponse( @@ -9091,7 +9754,9 @@ def test_list_usable_subnetworks_pages(transport_name: str = "grpc"): @pytest.mark.asyncio async def test_list_usable_subnetworks_async_pager(): - client = ClusterManagerAsyncClient(credentials=ga_credentials.AnonymousCredentials,) + client = ClusterManagerAsyncClient( + credentials=ga_credentials.AnonymousCredentials, + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -9110,10 +9775,13 @@ async def test_list_usable_subnetworks_async_pager(): next_page_token="abc", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[], next_page_token="def", + subnetworks=[], + next_page_token="def", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[cluster_service.UsableSubnetwork(),], + subnetworks=[ + cluster_service.UsableSubnetwork(), + ], next_page_token="ghi", ), cluster_service.ListUsableSubnetworksResponse( @@ -9124,10 +9792,12 @@ async def test_list_usable_subnetworks_async_pager(): ), RuntimeError, ) - async_pager = await client.list_usable_subnetworks(request={},) + async_pager = await client.list_usable_subnetworks( + request={}, + ) assert async_pager.next_page_token == "abc" responses = [] - async for response in async_pager: + async for response in async_pager: # pragma: no branch responses.append(response) assert len(responses) == 6 @@ -9136,7 +9806,9 @@ async def test_list_usable_subnetworks_async_pager(): @pytest.mark.asyncio async def test_list_usable_subnetworks_async_pages(): - client = ClusterManagerAsyncClient(credentials=ga_credentials.AnonymousCredentials,) + client = ClusterManagerAsyncClient( + credentials=ga_credentials.AnonymousCredentials, + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -9155,10 +9827,13 @@ async def test_list_usable_subnetworks_async_pages(): next_page_token="abc", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[], next_page_token="def", + subnetworks=[], + next_page_token="def", ), cluster_service.ListUsableSubnetworksResponse( - subnetworks=[cluster_service.UsableSubnetwork(),], + subnetworks=[ + cluster_service.UsableSubnetwork(), + ], next_page_token="ghi", ), cluster_service.ListUsableSubnetworksResponse( @@ -9170,16 +9845,25 @@ async def test_list_usable_subnetworks_async_pages(): RuntimeError, ) pages = [] - async for page_ in (await client.list_usable_subnetworks(request={})).pages: + async for page_ in ( + await client.list_usable_subnetworks(request={}) + ).pages: # pragma: no branch pages.append(page_) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token -@pytest.mark.parametrize("request_type", [cluster_service.ListLocationsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + cluster_service.ListLocationsRequest, + dict, + ], +) def test_list_locations(request_type, transport: str = "grpc"): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -9209,7 +9893,8 @@ def test_list_locations_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -9225,7 +9910,8 @@ async def test_list_locations_async( transport: str = "grpc_asyncio", request_type=cluster_service.ListLocationsRequest ): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -9258,13 +9944,15 @@ async def test_list_locations_async_from_dict(): def test_list_locations_field_headers(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cluster_service.ListLocationsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_locations), "__call__") as call: @@ -9278,7 +9966,10 @@ def test_list_locations_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -9291,7 +9982,7 @@ async def test_list_locations_field_headers_async(): # a field header. Set these to a non-empty value. request = cluster_service.ListLocationsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_locations), "__call__") as call: @@ -9307,11 +9998,16 @@ async def test_list_locations_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_locations_flattened(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_locations), "__call__") as call: @@ -9319,7 +10015,9 @@ def test_list_locations_flattened(): call.return_value = cluster_service.ListLocationsResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.list_locations(parent="parent_value",) + client.list_locations( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -9331,13 +10029,16 @@ def test_list_locations_flattened(): def test_list_locations_flattened_error(): - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.list_locations( - cluster_service.ListLocationsRequest(), parent="parent_value", + cluster_service.ListLocationsRequest(), + parent="parent_value", ) @@ -9357,7 +10058,9 @@ async def test_list_locations_flattened_async(): ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.list_locations(parent="parent_value",) + response = await client.list_locations( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -9378,7 +10081,8 @@ async def test_list_locations_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.list_locations( - cluster_service.ListLocationsRequest(), parent="parent_value", + cluster_service.ListLocationsRequest(), + parent="parent_value", ) @@ -9389,7 +10093,8 @@ def test_credentials_transport_error(): ) with pytest.raises(ValueError): client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # It is an error to provide a credentials file and a transport instance. @@ -9409,7 +10114,10 @@ def test_credentials_transport_error(): options = client_options.ClientOptions() options.api_key = "api_key" with pytest.raises(ValueError): - client = ClusterManagerClient(client_options=options, transport=transport,) + client = ClusterManagerClient( + client_options=options, + transport=transport, + ) # It is an error to provide an api_key and a credential. options = mock.Mock() @@ -9425,7 +10133,8 @@ def test_credentials_transport_error(): ) with pytest.raises(ValueError): client = ClusterManagerClient( - client_options={"scopes": ["1", "2"]}, transport=transport, + client_options={"scopes": ["1", "2"]}, + transport=transport, ) @@ -9468,10 +10177,28 @@ def test_transport_adc(transport_class): adc.assert_called_once() +@pytest.mark.parametrize( + "transport_name", + [ + "grpc", + ], +) +def test_transport_kind(transport_name): + transport = ClusterManagerClient.get_transport_class(transport_name)( + credentials=ga_credentials.AnonymousCredentials(), + ) + assert transport.kind == transport_name + + def test_transport_grpc_default(): # A client should use the gRPC transport by default. - client = ClusterManagerClient(credentials=ga_credentials.AnonymousCredentials(),) - assert isinstance(client.transport, transports.ClusterManagerGrpcTransport,) + client = ClusterManagerClient( + credentials=ga_credentials.AnonymousCredentials(), + ) + assert isinstance( + client.transport, + transports.ClusterManagerGrpcTransport, + ) def test_cluster_manager_base_transport_error(): @@ -9537,6 +10264,14 @@ def test_cluster_manager_base_transport(): with pytest.raises(NotImplementedError): transport.close() + # Catch all for all remaining methods and properties + remainder = [ + "kind", + ] + for r in remainder: + with pytest.raises(NotImplementedError): + getattr(transport, r)() + def test_cluster_manager_base_transport_with_credentials_file(): # Instantiate the base transport with a credentials file @@ -9548,7 +10283,8 @@ def test_cluster_manager_base_transport_with_credentials_file(): Transport.return_value = None load_creds.return_value = (ga_credentials.AnonymousCredentials(), None) transport = transports.ClusterManagerTransport( - credentials_file="credentials.json", quota_project_id="octopus", + credentials_file="credentials.json", + quota_project_id="octopus", ) load_creds.assert_called_once_with( "credentials.json", @@ -9681,24 +10417,40 @@ def test_cluster_manager_grpc_transport_client_cert_source_for_mtls(transport_cl ) -def test_cluster_manager_host_no_port(): +@pytest.mark.parametrize( + "transport_name", + [ + "grpc", + "grpc_asyncio", + ], +) +def test_cluster_manager_host_no_port(transport_name): client = ClusterManagerClient( credentials=ga_credentials.AnonymousCredentials(), client_options=client_options.ClientOptions( api_endpoint="container.googleapis.com" ), + transport=transport_name, ) - assert client.transport._host == "container.googleapis.com:443" + assert client.transport._host == ("container.googleapis.com:443") -def test_cluster_manager_host_with_port(): +@pytest.mark.parametrize( + "transport_name", + [ + "grpc", + "grpc_asyncio", + ], +) +def test_cluster_manager_host_with_port(transport_name): client = ClusterManagerClient( credentials=ga_credentials.AnonymousCredentials(), client_options=client_options.ClientOptions( api_endpoint="container.googleapis.com:8000" ), + transport=transport_name, ) - assert client.transport._host == "container.googleapis.com:8000" + assert client.transport._host == ("container.googleapis.com:8000") def test_cluster_manager_grpc_transport_channel(): @@ -9706,7 +10458,8 @@ def test_cluster_manager_grpc_transport_channel(): # Check that channel is used if provided. transport = transports.ClusterManagerGrpcTransport( - host="squid.clam.whelk", channel=channel, + host="squid.clam.whelk", + channel=channel, ) assert transport.grpc_channel == channel assert transport._host == "squid.clam.whelk:443" @@ -9718,7 +10471,8 @@ def test_cluster_manager_grpc_asyncio_transport_channel(): # Check that channel is used if provided. transport = transports.ClusterManagerGrpcAsyncIOTransport( - host="squid.clam.whelk", channel=channel, + host="squid.clam.whelk", + channel=channel, ) assert transport.grpc_channel == channel assert transport._host == "squid.clam.whelk:443" @@ -9828,7 +10582,10 @@ def test_cluster_manager_transport_channel_mtls_with_adc(transport_class): def test_topic_path(): project = "squid" topic = "clam" - expected = "projects/{project}/topics/{topic}".format(project=project, topic=topic,) + expected = "projects/{project}/topics/{topic}".format( + project=project, + topic=topic, + ) actual = ClusterManagerClient.topic_path(project, topic) assert expected == actual @@ -9867,7 +10624,9 @@ def test_parse_common_billing_account_path(): def test_common_folder_path(): folder = "cuttlefish" - expected = "folders/{folder}".format(folder=folder,) + expected = "folders/{folder}".format( + folder=folder, + ) actual = ClusterManagerClient.common_folder_path(folder) assert expected == actual @@ -9885,7 +10644,9 @@ def test_parse_common_folder_path(): def test_common_organization_path(): organization = "winkle" - expected = "organizations/{organization}".format(organization=organization,) + expected = "organizations/{organization}".format( + organization=organization, + ) actual = ClusterManagerClient.common_organization_path(organization) assert expected == actual @@ -9903,7 +10664,9 @@ def test_parse_common_organization_path(): def test_common_project_path(): project = "scallop" - expected = "projects/{project}".format(project=project,) + expected = "projects/{project}".format( + project=project, + ) actual = ClusterManagerClient.common_project_path(project) assert expected == actual @@ -9923,7 +10686,8 @@ def test_common_location_path(): project = "squid" location = "clam" expected = "projects/{project}/locations/{location}".format( - project=project, location=location, + project=project, + location=location, ) actual = ClusterManagerClient.common_location_path(project, location) assert expected == actual @@ -9948,7 +10712,8 @@ def test_client_with_default_client_info(): transports.ClusterManagerTransport, "_prep_wrapped_messages" ) as prep: client = ClusterManagerClient( - credentials=ga_credentials.AnonymousCredentials(), client_info=client_info, + credentials=ga_credentials.AnonymousCredentials(), + client_info=client_info, ) prep.assert_called_once_with(client_info) @@ -9957,7 +10722,8 @@ def test_client_with_default_client_info(): ) as prep: transport_class = ClusterManagerClient.get_transport_class() transport = transport_class( - credentials=ga_credentials.AnonymousCredentials(), client_info=client_info, + credentials=ga_credentials.AnonymousCredentials(), + client_info=client_info, ) prep.assert_called_once_with(client_info) @@ -9965,7 +10731,8 @@ def test_client_with_default_client_info(): @pytest.mark.asyncio async def test_transport_close_async(): client = ClusterManagerAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc_asyncio", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc_asyncio", ) with mock.patch.object( type(getattr(client.transport, "grpc_channel")), "close"