diff --git a/docs/conf.py b/docs/conf.py index a2df03273..e6a3d0d1a 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -346,7 +346,10 @@ intersphinx_mapping = { "python": ("http://python.readthedocs.org/en/latest/", None), "google-auth": ("https://google-auth.readthedocs.io/en/stable", None), - "google.api_core": ("https://googleapis.dev/python/google-api-core/latest/", None,), + "google.api_core": ( + "https://googleapis.dev/python/google-api-core/latest/", + None, + ), "grpc": ("https://grpc.io/grpc/python/", None), } diff --git a/google/cloud/bigtable/backup.py b/google/cloud/bigtable/backup.py index c6a2826dd..03a1c894e 100644 --- a/google/cloud/bigtable/backup.py +++ b/google/cloud/bigtable/backup.py @@ -208,7 +208,7 @@ def size_bytes(self): @property def state(self): - """ The current state of this Backup. + """The current state of this Backup. :rtype: :class:`~google.cloud.bigtable_admin_v2.gapic.enums.Backup.State` :returns: The current state of this Backup. @@ -358,7 +358,8 @@ def update_expire_time(self, new_expire_time): :param new_expire_time: the new expiration time timestamp """ backup_update = table_pb2.Backup( - name=self.name, expire_time=_datetime_to_pb_timestamp(new_expire_time), + name=self.name, + expire_time=_datetime_to_pb_timestamp(new_expire_time), ) update_mask = field_mask_pb2.FieldMask(paths=["expire_time"]) api = self._instance._client.table_admin_client diff --git a/google/cloud/bigtable/batcher.py b/google/cloud/bigtable/batcher.py index 0994e289d..782cb979a 100644 --- a/google/cloud/bigtable/batcher.py +++ b/google/cloud/bigtable/batcher.py @@ -25,7 +25,7 @@ class MaxMutationsError(ValueError): class MutationsBatcher(object): - """ A MutationsBatcher is used in batch cases where the number of mutations + """A MutationsBatcher is used in batch cases where the number of mutations is large or unknown. It will store DirectRows in memory until one of the size limits is reached, or an explicit call to flush() is performed. When a flush event occurs, the DirectRows in memory will be sent to Cloud @@ -65,7 +65,7 @@ def __init__(self, table, flush_count=FLUSH_COUNT, max_row_bytes=MAX_ROW_BYTES): self.max_row_bytes = max_row_bytes def mutate(self, row): - """ Add a row to the batch. If the current batch meets one of the size + """Add a row to the batch. If the current batch meets one of the size limits, the batch is sent synchronously. For example: @@ -105,7 +105,7 @@ def mutate(self, row): self.flush() def mutate_rows(self, rows): - """ Add a row to the batch. If the current batch meets one of the size + """Add a row to the batch. If the current batch meets one of the size limits, the batch is sent synchronously. For example: @@ -130,7 +130,7 @@ def mutate_rows(self, rows): self.mutate(row) def flush(self): - """ Sends the current. batch to Cloud Bigtable. + """Sends the current. batch to Cloud Bigtable. For example: .. literalinclude:: snippets.py diff --git a/google/cloud/bigtable/client.py b/google/cloud/bigtable/client.py index bbb830519..2ee6e7c77 100644 --- a/google/cloud/bigtable/client.py +++ b/google/cloud/bigtable/client.py @@ -175,7 +175,9 @@ def __init__( self._channel = channel self.SCOPE = self._get_scopes() super(Client, self).__init__( - project=project, credentials=credentials, client_options=client_options, + project=project, + credentials=credentials, + client_options=client_options, ) def _get_scopes(self): diff --git a/google/cloud/bigtable/row.py b/google/cloud/bigtable/row.py index 92f5b818b..b28b86aa2 100644 --- a/google/cloud/bigtable/row.py +++ b/google/cloud/bigtable/row.py @@ -295,7 +295,7 @@ def _get_mutations(self, state=None): # pylint: disable=unused-argument return self._pb_mutations def get_mutations_size(self): - """ Gets the total mutations size for current row + """Gets the total mutations size for current row For example: diff --git a/google/cloud/bigtable/row_data.py b/google/cloud/bigtable/row_data.py index 38bf85956..04824e1be 100644 --- a/google/cloud/bigtable/row_data.py +++ b/google/cloud/bigtable/row_data.py @@ -595,7 +595,7 @@ def _copy_from_previous(self, cell): class _ReadRowsRequestManager(object): - """ Update the ReadRowsRequest message in case of failures by + """Update the ReadRowsRequest message in case of failures by filtering the already read keys. :type message: class:`data_messages_v2_pb2.ReadRowsRequest` @@ -617,8 +617,7 @@ def __init__(self, message, last_scanned_key, rows_read_so_far): self.rows_read_so_far = rows_read_so_far def build_updated_request(self): - """ Updates the given message request as per last scanned key - """ + """Updates the given message request as per last scanned key""" r_kwargs = { "table_name": self.message.table_name, "filter": self.message.filter, diff --git a/google/cloud/bigtable/row_set.py b/google/cloud/bigtable/row_set.py index 5de7dabff..e229c805a 100644 --- a/google/cloud/bigtable/row_set.py +++ b/google/cloud/bigtable/row_set.py @@ -19,10 +19,10 @@ class RowSet(object): - """ Convenience wrapper of google.bigtable.v2.RowSet + """Convenience wrapper of google.bigtable.v2.RowSet - Useful for creating a set of row keys and row ranges, which can - be passed to yield_rows method of class:`.Table.yield_rows`. + Useful for creating a set of row keys and row ranges, which can + be passed to yield_rows method of class:`.Table.yield_rows`. """ def __init__(self): @@ -145,7 +145,7 @@ def _update_message_request(self, message): class RowRange(object): - """ Convenience wrapper of google.bigtable.v2.RowRange + """Convenience wrapper of google.bigtable.v2.RowRange :type start_key: bytes :param start_key: (Optional) Start key of the row range. If left empty, @@ -195,7 +195,7 @@ def __ne__(self, other): return not self == other def get_range_kwargs(self): - """ Convert row range object to dict which can be passed to + """Convert row range object to dict which can be passed to google.bigtable.v2.RowRange add method. """ range_kwargs = {} diff --git a/google/cloud/bigtable_admin_v2/gapic/bigtable_instance_admin_client.py b/google/cloud/bigtable_admin_v2/gapic/bigtable_instance_admin_client.py index 8edb3c168..d27154d53 100644 --- a/google/cloud/bigtable_admin_v2/gapic/bigtable_instance_admin_client.py +++ b/google/cloud/bigtable_admin_v2/gapic/bigtable_instance_admin_client.py @@ -130,7 +130,8 @@ def location_path(cls, project, location): def project_path(cls, project): """Return a fully-qualified project string.""" return google.api_core.path_template.expand( - "projects/{project}", project=project, + "projects/{project}", + project=project, ) def __init__( @@ -220,7 +221,9 @@ def __init__( self.transport = transport else: self.transport = bigtable_instance_admin_grpc_transport.BigtableInstanceAdminGrpcTransport( - address=api_endpoint, channel=channel, credentials=credentials, + address=api_endpoint, + channel=channel, + credentials=credentials, ) if client_info is None: @@ -416,7 +419,9 @@ def get_instance( client_info=self._client_info, ) - request = bigtable_instance_admin_pb2.GetInstanceRequest(name=name,) + request = bigtable_instance_admin_pb2.GetInstanceRequest( + name=name, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -489,7 +494,8 @@ def list_instances( ) request = bigtable_instance_admin_pb2.ListInstancesRequest( - parent=parent, page_token=page_token, + parent=parent, + page_token=page_token, ) if metadata is None: metadata = [] @@ -683,7 +689,8 @@ def partial_update_instance( ) request = bigtable_instance_admin_pb2.PartialUpdateInstanceRequest( - instance=instance, update_mask=update_mask, + instance=instance, + update_mask=update_mask, ) if metadata is None: metadata = [] @@ -757,7 +764,9 @@ def delete_instance( client_info=self._client_info, ) - request = bigtable_instance_admin_pb2.DeleteInstanceRequest(name=name,) + request = bigtable_instance_admin_pb2.DeleteInstanceRequest( + name=name, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -854,7 +863,9 @@ def create_cluster( ) request = bigtable_instance_admin_pb2.CreateClusterRequest( - parent=parent, cluster_id=cluster_id, cluster=cluster, + parent=parent, + cluster_id=cluster_id, + cluster=cluster, ) if metadata is None: metadata = [] @@ -931,7 +942,9 @@ def get_cluster( client_info=self._client_info, ) - request = bigtable_instance_admin_pb2.GetClusterRequest(name=name,) + request = bigtable_instance_admin_pb2.GetClusterRequest( + name=name, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -1007,7 +1020,8 @@ def list_clusters( ) request = bigtable_instance_admin_pb2.ListClustersRequest( - parent=parent, page_token=page_token, + parent=parent, + page_token=page_token, ) if metadata is None: metadata = [] @@ -1181,7 +1195,9 @@ def delete_cluster( client_info=self._client_info, ) - request = bigtable_instance_admin_pb2.DeleteClusterRequest(name=name,) + request = bigtable_instance_admin_pb2.DeleteClusterRequest( + name=name, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -1346,7 +1362,9 @@ def get_app_profile( client_info=self._client_info, ) - request = bigtable_instance_admin_pb2.GetAppProfileRequest(name=name,) + request = bigtable_instance_admin_pb2.GetAppProfileRequest( + name=name, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -1441,7 +1459,8 @@ def list_app_profiles( ) request = bigtable_instance_admin_pb2.ListAppProfilesRequest( - parent=parent, page_size=page_size, + parent=parent, + page_size=page_size, ) if metadata is None: metadata = [] @@ -1627,7 +1646,8 @@ def delete_app_profile( ) request = bigtable_instance_admin_pb2.DeleteAppProfileRequest( - name=name, ignore_warnings=ignore_warnings, + name=name, + ignore_warnings=ignore_warnings, ) if metadata is None: metadata = [] @@ -1707,7 +1727,8 @@ def get_iam_policy( ) request = iam_policy_pb2.GetIamPolicyRequest( - resource=resource, options=options_, + resource=resource, + options=options_, ) if metadata is None: metadata = [] @@ -1791,7 +1812,10 @@ def set_iam_policy( client_info=self._client_info, ) - request = iam_policy_pb2.SetIamPolicyRequest(resource=resource, policy=policy,) + request = iam_policy_pb2.SetIamPolicyRequest( + resource=resource, + policy=policy, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -1871,7 +1895,8 @@ def test_iam_permissions( ) request = iam_policy_pb2.TestIamPermissionsRequest( - resource=resource, permissions=permissions, + resource=resource, + permissions=permissions, ) if metadata is None: metadata = [] diff --git a/google/cloud/bigtable_admin_v2/gapic/bigtable_table_admin_client.py b/google/cloud/bigtable_admin_v2/gapic/bigtable_table_admin_client.py index cac517314..acbc4b26f 100644 --- a/google/cloud/bigtable_admin_v2/gapic/bigtable_table_admin_client.py +++ b/google/cloud/bigtable_admin_v2/gapic/bigtable_table_admin_client.py @@ -232,8 +232,12 @@ def __init__( ) self.transport = transport else: - self.transport = bigtable_table_admin_grpc_transport.BigtableTableAdminGrpcTransport( - address=api_endpoint, channel=channel, credentials=credentials, + self.transport = ( + bigtable_table_admin_grpc_transport.BigtableTableAdminGrpcTransport( + address=api_endpoint, + channel=channel, + credentials=credentials, + ) ) if client_info is None: @@ -457,7 +461,9 @@ def create_table_from_snapshot( ) request = bigtable_table_admin_pb2.CreateTableFromSnapshotRequest( - parent=parent, table_id=table_id, source_snapshot=source_snapshot, + parent=parent, + table_id=table_id, + source_snapshot=source_snapshot, ) if metadata is None: metadata = [] @@ -560,7 +566,9 @@ def list_tables( ) request = bigtable_table_admin_pb2.ListTablesRequest( - parent=parent, view=view, page_size=page_size, + parent=parent, + view=view, + page_size=page_size, ) if metadata is None: metadata = [] @@ -645,7 +653,10 @@ def get_table( client_info=self._client_info, ) - request = bigtable_table_admin_pb2.GetTableRequest(name=name, view=view,) + request = bigtable_table_admin_pb2.GetTableRequest( + name=name, + view=view, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -712,7 +723,9 @@ def delete_table( client_info=self._client_info, ) - request = bigtable_table_admin_pb2.DeleteTableRequest(name=name,) + request = bigtable_table_admin_pb2.DeleteTableRequest( + name=name, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -798,7 +811,8 @@ def modify_column_families( ) request = bigtable_table_admin_pb2.ModifyColumnFamiliesRequest( - name=name, modifications=modifications, + name=name, + modifications=modifications, ) if metadata is None: metadata = [] @@ -961,7 +975,9 @@ def generate_consistency_token( client_info=self._client_info, ) - request = bigtable_table_admin_pb2.GenerateConsistencyTokenRequest(name=name,) + request = bigtable_table_admin_pb2.GenerateConsistencyTokenRequest( + name=name, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -1040,7 +1056,8 @@ def check_consistency( ) request = bigtable_table_admin_pb2.CheckConsistencyRequest( - name=name, consistency_token=consistency_token, + name=name, + consistency_token=consistency_token, ) if metadata is None: metadata = [] @@ -1121,7 +1138,8 @@ def get_iam_policy( ) request = iam_policy_pb2.GetIamPolicyRequest( - resource=resource, options=options_, + resource=resource, + options=options_, ) if metadata is None: metadata = [] @@ -1205,7 +1223,10 @@ def set_iam_policy( client_info=self._client_info, ) - request = iam_policy_pb2.SetIamPolicyRequest(resource=resource, policy=policy,) + request = iam_policy_pb2.SetIamPolicyRequest( + resource=resource, + policy=policy, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -1285,7 +1306,8 @@ def test_iam_permissions( ) request = iam_policy_pb2.TestIamPermissionsRequest( - resource=resource, permissions=permissions, + resource=resource, + permissions=permissions, ) if metadata is None: metadata = [] @@ -1485,7 +1507,9 @@ def get_snapshot( client_info=self._client_info, ) - request = bigtable_table_admin_pb2.GetSnapshotRequest(name=name,) + request = bigtable_table_admin_pb2.GetSnapshotRequest( + name=name, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -1586,7 +1610,8 @@ def list_snapshots( ) request = bigtable_table_admin_pb2.ListSnapshotsRequest( - parent=parent, page_size=page_size, + parent=parent, + page_size=page_size, ) if metadata is None: metadata = [] @@ -1672,7 +1697,9 @@ def delete_snapshot( client_info=self._client_info, ) - request = bigtable_table_admin_pb2.DeleteSnapshotRequest(name=name,) + request = bigtable_table_admin_pb2.DeleteSnapshotRequest( + name=name, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -1777,7 +1804,9 @@ def create_backup( ) request = bigtable_table_admin_pb2.CreateBackupRequest( - parent=parent, backup_id=backup_id, backup=backup, + parent=parent, + backup_id=backup_id, + backup=backup, ) if metadata is None: metadata = [] @@ -1854,7 +1883,9 @@ def get_backup( client_info=self._client_info, ) - request = bigtable_table_admin_pb2.GetBackupRequest(name=name,) + request = bigtable_table_admin_pb2.GetBackupRequest( + name=name, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -2000,7 +2031,10 @@ def list_backups( ) request = bigtable_table_admin_pb2.ListBackupsRequest( - parent=parent, filter=filter_, order_by=order_by, page_size=page_size, + parent=parent, + filter=filter_, + order_by=order_by, + page_size=page_size, ) if metadata is None: metadata = [] @@ -2102,7 +2136,8 @@ def update_backup( ) request = bigtable_table_admin_pb2.UpdateBackupRequest( - backup=backup, update_mask=update_mask, + backup=backup, + update_mask=update_mask, ) if metadata is None: metadata = [] @@ -2170,7 +2205,9 @@ def delete_backup( client_info=self._client_info, ) - request = bigtable_table_admin_pb2.DeleteBackupRequest(name=name,) + request = bigtable_table_admin_pb2.DeleteBackupRequest( + name=name, + ) if metadata is None: metadata = [] metadata = list(metadata) @@ -2263,10 +2300,14 @@ def restore_table( # Sanity check: We have some fields which are mutually exclusive; # raise ValueError if more than one is sent. - google.api_core.protobuf_helpers.check_oneof(backup=backup,) + google.api_core.protobuf_helpers.check_oneof( + backup=backup, + ) request = bigtable_table_admin_pb2.RestoreTableRequest( - parent=parent, table_id=table_id, backup=backup, + parent=parent, + table_id=table_id, + backup=backup, ) if metadata is None: metadata = [] diff --git a/google/cloud/bigtable_admin_v2/proto/bigtable_instance_admin_pb2.py b/google/cloud/bigtable_admin_v2/proto/bigtable_instance_admin_pb2.py index 63590907a..38fe53f88 100644 --- a/google/cloud/bigtable_admin_v2/proto/bigtable_instance_admin_pb2.py +++ b/google/cloud/bigtable_admin_v2/proto/bigtable_instance_admin_pb2.py @@ -197,7 +197,9 @@ ), ], extensions=[], - nested_types=[_CREATEINSTANCEREQUEST_CLUSTERSENTRY,], + nested_types=[ + _CREATEINSTANCEREQUEST_CLUSTERSENTRY, + ], enum_types=[], serialized_options=None, is_extendable=False, diff --git a/google/cloud/bigtable_admin_v2/proto/bigtable_instance_admin_pb2_grpc.py b/google/cloud/bigtable_admin_v2/proto/bigtable_instance_admin_pb2_grpc.py index 8b1395579..0337e5d4f 100644 --- a/google/cloud/bigtable_admin_v2/proto/bigtable_instance_admin_pb2_grpc.py +++ b/google/cloud/bigtable_admin_v2/proto/bigtable_instance_admin_pb2_grpc.py @@ -132,22 +132,19 @@ class BigtableInstanceAdminServicer(object): """ def CreateInstance(self, request, context): - """Create an instance within a project. - """ + """Create an instance within a project.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def GetInstance(self, request, context): - """Gets information about an instance. - """ + """Gets information about an instance.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def ListInstances(self, request, context): - """Lists information about instances in a project. - """ + """Lists information about instances in a project.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") @@ -170,78 +167,67 @@ def PartialUpdateInstance(self, request, context): raise NotImplementedError("Method not implemented!") def DeleteInstance(self, request, context): - """Delete an instance from a project. - """ + """Delete an instance from a project.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def CreateCluster(self, request, context): - """Creates a cluster within an instance. - """ + """Creates a cluster within an instance.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def GetCluster(self, request, context): - """Gets information about a cluster. - """ + """Gets information about a cluster.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def ListClusters(self, request, context): - """Lists information about clusters in an instance. - """ + """Lists information about clusters in an instance.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def UpdateCluster(self, request, context): - """Updates a cluster within an instance. - """ + """Updates a cluster within an instance.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def DeleteCluster(self, request, context): - """Deletes a cluster from an instance. - """ + """Deletes a cluster from an instance.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def CreateAppProfile(self, request, context): - """Creates an app profile within an instance. - """ + """Creates an app profile within an instance.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def GetAppProfile(self, request, context): - """Gets information about an app profile. - """ + """Gets information about an app profile.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def ListAppProfiles(self, request, context): - """Lists information about app profiles in an instance. - """ + """Lists information about app profiles in an instance.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def UpdateAppProfile(self, request, context): - """Updates an app profile within an instance. - """ + """Updates an app profile within an instance.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def DeleteAppProfile(self, request, context): - """Deletes an app profile from an instance. - """ + """Deletes an app profile from an instance.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") @@ -263,8 +249,7 @@ def SetIamPolicy(self, request, context): raise NotImplementedError("Method not implemented!") def TestIamPermissions(self, request, context): - """Returns permissions that the caller has on the specified instance resource. - """ + """Returns permissions that the caller has on the specified instance resource.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") diff --git a/google/cloud/bigtable_admin_v2/proto/bigtable_table_admin_pb2.py b/google/cloud/bigtable_admin_v2/proto/bigtable_table_admin_pb2.py index 5ca167d87..c7094eac2 100644 --- a/google/cloud/bigtable_admin_v2/proto/bigtable_table_admin_pb2.py +++ b/google/cloud/bigtable_admin_v2/proto/bigtable_table_admin_pb2.py @@ -184,7 +184,9 @@ ), ], extensions=[], - nested_types=[_CREATETABLEREQUEST_SPLIT,], + nested_types=[ + _CREATETABLEREQUEST_SPLIT, + ], enum_types=[], serialized_options=None, is_extendable=False, @@ -776,7 +778,9 @@ ), ], extensions=[], - nested_types=[_MODIFYCOLUMNFAMILIESREQUEST_MODIFICATION,], + nested_types=[ + _MODIFYCOLUMNFAMILIESREQUEST_MODIFICATION, + ], enum_types=[], serialized_options=None, is_extendable=False, diff --git a/google/cloud/bigtable_admin_v2/proto/bigtable_table_admin_pb2_grpc.py b/google/cloud/bigtable_admin_v2/proto/bigtable_table_admin_pb2_grpc.py index 2b8d46e20..949de429e 100644 --- a/google/cloud/bigtable_admin_v2/proto/bigtable_table_admin_pb2_grpc.py +++ b/google/cloud/bigtable_admin_v2/proto/bigtable_table_admin_pb2_grpc.py @@ -174,22 +174,19 @@ def CreateTableFromSnapshot(self, request, context): raise NotImplementedError("Method not implemented!") def ListTables(self, request, context): - """Lists all tables served from a specified instance. - """ + """Lists all tables served from a specified instance.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def GetTable(self, request, context): - """Gets metadata information about the specified table. - """ + """Gets metadata information about the specified table.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def DeleteTable(self, request, context): - """Permanently deletes a specified table and all of its data. - """ + """Permanently deletes a specified table and all of its data.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") @@ -300,22 +297,19 @@ def CreateBackup(self, request, context): raise NotImplementedError("Method not implemented!") def GetBackup(self, request, context): - """Gets metadata on a pending or completed Cloud Bigtable Backup. - """ + """Gets metadata on a pending or completed Cloud Bigtable Backup.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def UpdateBackup(self, request, context): - """Updates a pending or completed Cloud Bigtable Backup. - """ + """Updates a pending or completed Cloud Bigtable Backup.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") def DeleteBackup(self, request, context): - """Deletes a pending or completed Cloud Bigtable backup. - """ + """Deletes a pending or completed Cloud Bigtable backup.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") @@ -360,8 +354,7 @@ def SetIamPolicy(self, request, context): raise NotImplementedError("Method not implemented!") def TestIamPermissions(self, request, context): - """Returns permissions that the caller has on the specified table resource. - """ + """Returns permissions that the caller has on the specified table resource.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") diff --git a/google/cloud/bigtable_admin_v2/proto/common_pb2.py b/google/cloud/bigtable_admin_v2/proto/common_pb2.py index 09233cff5..e07dea1d1 100644 --- a/google/cloud/bigtable_admin_v2/proto/common_pb2.py +++ b/google/cloud/bigtable_admin_v2/proto/common_pb2.py @@ -23,7 +23,9 @@ serialized_options=b'\n\034com.google.bigtable.admin.v2B\013CommonProtoP\001Z=google.golang.org/genproto/googleapis/bigtable/admin/v2;admin\252\002\036Google.Cloud.Bigtable.Admin.V2\312\002\036Google\\Cloud\\Bigtable\\Admin\\V2\352\002"Google::Cloud::Bigtable::Admin::V2', create_key=_descriptor._internal_create_key, serialized_pb=b'\n1google/cloud/bigtable_admin_v2/proto/common.proto\x12\x18google.bigtable.admin.v2\x1a\x1fgoogle/protobuf/timestamp.proto"\x8b\x01\n\x11OperationProgress\x12\x18\n\x10progress_percent\x18\x01 \x01(\x05\x12.\n\nstart_time\x18\x02 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12,\n\x08\x65nd_time\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.Timestamp*=\n\x0bStorageType\x12\x1c\n\x18STORAGE_TYPE_UNSPECIFIED\x10\x00\x12\x07\n\x03SSD\x10\x01\x12\x07\n\x03HDD\x10\x02\x42\xd3\x01\n\x1c\x63om.google.bigtable.admin.v2B\x0b\x43ommonProtoP\x01Z=google.golang.org/genproto/googleapis/bigtable/admin/v2;admin\xaa\x02\x1eGoogle.Cloud.Bigtable.Admin.V2\xca\x02\x1eGoogle\\Cloud\\Bigtable\\Admin\\V2\xea\x02"Google::Cloud::Bigtable::Admin::V2b\x06proto3', - dependencies=[google_dot_protobuf_dot_timestamp__pb2.DESCRIPTOR,], + dependencies=[ + google_dot_protobuf_dot_timestamp__pb2.DESCRIPTOR, + ], ) _STORAGETYPE = _descriptor.EnumDescriptor( diff --git a/google/cloud/bigtable_admin_v2/proto/instance_pb2.py b/google/cloud/bigtable_admin_v2/proto/instance_pb2.py index e0138e0fb..4f3ce0a5b 100644 --- a/google/cloud/bigtable_admin_v2/proto/instance_pb2.py +++ b/google/cloud/bigtable_admin_v2/proto/instance_pb2.py @@ -332,8 +332,13 @@ ), ], extensions=[], - nested_types=[_INSTANCE_LABELSENTRY,], - enum_types=[_INSTANCE_STATE, _INSTANCE_TYPE,], + nested_types=[ + _INSTANCE_LABELSENTRY, + ], + enum_types=[ + _INSTANCE_STATE, + _INSTANCE_TYPE, + ], serialized_options=b"\352AK\n bigtable.googleapis.com/Instance\022'projects/{project}/instances/{instance}", is_extendable=False, syntax="proto3", @@ -450,7 +455,9 @@ ], extensions=[], nested_types=[], - enum_types=[_CLUSTER_STATE,], + enum_types=[ + _CLUSTER_STATE, + ], serialized_options=b"\352A]\n\037bigtable.googleapis.com/Cluster\022:projects/{project}/instances/{instance}/clusters/{cluster}", is_extendable=False, syntax="proto3", diff --git a/google/cloud/bigtable_admin_v2/proto/table_pb2.py b/google/cloud/bigtable_admin_v2/proto/table_pb2.py index 67238a81e..71191acba 100644 --- a/google/cloud/bigtable_admin_v2/proto/table_pb2.py +++ b/google/cloud/bigtable_admin_v2/proto/table_pb2.py @@ -397,7 +397,9 @@ ], extensions=[], nested_types=[], - enum_types=[_TABLE_CLUSTERSTATE_REPLICATIONSTATE,], + enum_types=[ + _TABLE_CLUSTERSTATE_REPLICATIONSTATE, + ], serialized_options=None, is_extendable=False, syntax="proto3", @@ -635,7 +637,10 @@ _TABLE_CLUSTERSTATESENTRY, _TABLE_COLUMNFAMILIESENTRY, ], - enum_types=[_TABLE_TIMESTAMPGRANULARITY, _TABLE_VIEW,], + enum_types=[ + _TABLE_TIMESTAMPGRANULARITY, + _TABLE_VIEW, + ], serialized_options=b"\352AW\n\035bigtable.googleapis.com/Table\0226projects/{project}/instances/{instance}/tables/{table}", is_extendable=False, syntax="proto3", @@ -853,7 +858,10 @@ ), ], extensions=[], - nested_types=[_GCRULE_INTERSECTION, _GCRULE_UNION,], + nested_types=[ + _GCRULE_INTERSECTION, + _GCRULE_UNION, + ], enum_types=[], serialized_options=None, is_extendable=False, @@ -1018,7 +1026,9 @@ ], extensions=[], nested_types=[], - enum_types=[_SNAPSHOT_STATE,], + enum_types=[ + _SNAPSHOT_STATE, + ], serialized_options=b"\352As\n bigtable.googleapis.com/Snapshot\022Oprojects/{project}/instances/{instance}/clusters/{cluster}/snapshots/{snapshot}", is_extendable=False, syntax="proto3", @@ -1173,7 +1183,9 @@ ], extensions=[], nested_types=[], - enum_types=[_BACKUP_STATE,], + enum_types=[ + _BACKUP_STATE, + ], serialized_options=b"\352Am\n\036bigtable.googleapis.com/Backup\022Kprojects/{project}/instances/{instance}/clusters/{cluster}/backups/{backup}", is_extendable=False, syntax="proto3", diff --git a/google/cloud/bigtable_v2/gapic/bigtable_client.py b/google/cloud/bigtable_v2/gapic/bigtable_client.py index f02e0048f..7e544c99e 100644 --- a/google/cloud/bigtable_v2/gapic/bigtable_client.py +++ b/google/cloud/bigtable_v2/gapic/bigtable_client.py @@ -168,7 +168,9 @@ def __init__( self.transport = transport else: self.transport = bigtable_grpc_transport.BigtableGrpcTransport( - address=api_endpoint, channel=channel, credentials=credentials, + address=api_endpoint, + channel=channel, + credentials=credentials, ) if client_info is None: @@ -356,7 +358,8 @@ def sample_row_keys( ) request = bigtable_pb2.SampleRowKeysRequest( - table_name=table_name, app_profile_id=app_profile_id, + table_name=table_name, + app_profile_id=app_profile_id, ) if metadata is None: metadata = [] @@ -541,7 +544,9 @@ def mutate_rows( ) request = bigtable_pb2.MutateRowsRequest( - table_name=table_name, entries=entries, app_profile_id=app_profile_id, + table_name=table_name, + entries=entries, + app_profile_id=app_profile_id, ) if metadata is None: metadata = [] diff --git a/google/cloud/bigtable_v2/proto/bigtable_pb2.py b/google/cloud/bigtable_v2/proto/bigtable_pb2.py index ba711b20c..f6d825d89 100644 --- a/google/cloud/bigtable_v2/proto/bigtable_pb2.py +++ b/google/cloud/bigtable_v2/proto/bigtable_pb2.py @@ -408,7 +408,9 @@ ), ], extensions=[], - nested_types=[_READROWSRESPONSE_CELLCHUNK,], + nested_types=[ + _READROWSRESPONSE_CELLCHUNK, + ], enum_types=[], serialized_options=None, is_extendable=False, @@ -785,7 +787,9 @@ ), ], extensions=[], - nested_types=[_MUTATEROWSREQUEST_ENTRY,], + nested_types=[ + _MUTATEROWSREQUEST_ENTRY, + ], enum_types=[], serialized_options=None, is_extendable=False, @@ -885,7 +889,9 @@ ), ], extensions=[], - nested_types=[_MUTATEROWSRESPONSE_ENTRY,], + nested_types=[ + _MUTATEROWSRESPONSE_ENTRY, + ], enum_types=[], serialized_options=None, is_extendable=False, diff --git a/google/cloud/bigtable_v2/proto/bigtable_pb2_grpc.py b/google/cloud/bigtable_v2/proto/bigtable_pb2_grpc.py index db4ee99f3..2a094a7f9 100644 --- a/google/cloud/bigtable_v2/proto/bigtable_pb2_grpc.py +++ b/google/cloud/bigtable_v2/proto/bigtable_pb2_grpc.py @@ -8,8 +8,7 @@ class BigtableStub(object): - """Service for reading from and writing to existing Bigtable tables. - """ + """Service for reading from and writing to existing Bigtable tables.""" def __init__(self, channel): """Constructor. @@ -50,8 +49,7 @@ def __init__(self, channel): class BigtableServicer(object): - """Service for reading from and writing to existing Bigtable tables. - """ + """Service for reading from and writing to existing Bigtable tables.""" def ReadRows(self, request, context): """Streams back the contents of all requested rows in key order, optionally @@ -92,8 +90,7 @@ def MutateRows(self, request, context): raise NotImplementedError("Method not implemented!") def CheckAndMutateRow(self, request, context): - """Mutates a row atomically based on the output of a predicate Reader filter. - """ + """Mutates a row atomically based on the output of a predicate Reader filter.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details("Method not implemented!") raise NotImplementedError("Method not implemented!") @@ -151,8 +148,7 @@ def add_BigtableServicer_to_server(servicer, server): # This class is part of an EXPERIMENTAL API. class Bigtable(object): - """Service for reading from and writing to existing Bigtable tables. - """ + """Service for reading from and writing to existing Bigtable tables.""" @staticmethod def ReadRows( diff --git a/google/cloud/bigtable_v2/proto/data_pb2.py b/google/cloud/bigtable_v2/proto/data_pb2.py index a64f9b10e..5f62756a8 100644 --- a/google/cloud/bigtable_v2/proto/data_pb2.py +++ b/google/cloud/bigtable_v2/proto/data_pb2.py @@ -1294,7 +1294,11 @@ ), ], extensions=[], - nested_types=[_ROWFILTER_CHAIN, _ROWFILTER_INTERLEAVE, _ROWFILTER_CONDITION,], + nested_types=[ + _ROWFILTER_CHAIN, + _ROWFILTER_INTERLEAVE, + _ROWFILTER_CONDITION, + ], enum_types=[], serialized_options=None, is_extendable=False, diff --git a/synth.metadata b/synth.metadata index a4d60cb48..79b395fba 100644 --- a/synth.metadata +++ b/synth.metadata @@ -4,7 +4,7 @@ "git": { "name": ".", "remote": "https://github.com/googleapis/python-bigtable.git", - "sha": "3ecca7a7b52b0f4fc38db5c5016622b994c1a8aa" + "sha": "dfe658a2b1270eda7a8a084aca28d65b3297a04f" } }, { diff --git a/tests/unit/test_backup.py b/tests/unit/test_backup.py index 587202a84..2f263dffd 100644 --- a/tests/unit/test_backup.py +++ b/tests/unit/test_backup.py @@ -186,7 +186,10 @@ def test_property_cluster_setter(self): self.assertEqual(backup.cluster, self.CLUSTER_ID) def test_property_parent_none(self): - backup = self._make_one(self.BACKUP_ID, _Instance(self.INSTANCE_NAME),) + backup = self._make_one( + self.BACKUP_ID, + _Instance(self.INSTANCE_NAME), + ) self.assertIsNone(backup.parent) def test_property_parent_w_cluster(self): @@ -318,7 +321,9 @@ def test_create_grpc_error(self): backup.create(self.CLUSTER_ID) api.create_backup.assert_called_once_with( - parent=self.CLUSTER_NAME, backup_id=self.BACKUP_ID, backup=backup_pb, + parent=self.CLUSTER_NAME, + backup_id=self.BACKUP_ID, + backup=backup_pb, ) def test_create_already_exists(self): @@ -347,7 +352,9 @@ def test_create_already_exists(self): backup.create(self.CLUSTER_ID) api.create_backup.assert_called_once_with( - parent=self.CLUSTER_NAME, backup_id=self.BACKUP_ID, backup=backup_pb, + parent=self.CLUSTER_NAME, + backup_id=self.BACKUP_ID, + backup=backup_pb, ) def test_create_instance_not_found(self): @@ -376,7 +383,9 @@ def test_create_instance_not_found(self): backup.create(self.CLUSTER_ID) api.create_backup.assert_called_once_with( - parent=self.CLUSTER_NAME, backup_id=self.BACKUP_ID, backup=backup_pb, + parent=self.CLUSTER_NAME, + backup_id=self.BACKUP_ID, + backup=backup_pb, ) def test_create_cluster_not_set(self): @@ -402,7 +411,9 @@ def test_create_table_not_set(self): def test_create_expire_time_not_set(self): backup = self._make_one( - self.BACKUP_ID, _Instance(self.INSTANCE_NAME), table_id=self.TABLE_ID, + self.BACKUP_ID, + _Instance(self.INSTANCE_NAME), + table_id=self.TABLE_ID, ) with self.assertRaises(ValueError): @@ -435,7 +446,9 @@ def test_create_success(self): self.assertIs(future, op_future) api.create_backup.assert_called_once_with( - parent=self.CLUSTER_NAME, backup_id=self.BACKUP_ID, backup=backup_pb, + parent=self.CLUSTER_NAME, + backup_id=self.BACKUP_ID, + backup=backup_pb, ) def test_exists_grpc_error(self): @@ -598,11 +611,13 @@ def test_update_expire_time_grpc_error(self): backup.update_expire_time(expire_time) backup_update = table_pb2.Backup( - name=self.BACKUP_NAME, expire_time=_datetime_to_pb_timestamp(expire_time), + name=self.BACKUP_NAME, + expire_time=_datetime_to_pb_timestamp(expire_time), ) update_mask = field_mask_pb2.FieldMask(paths=["expire_time"]) api.update_backup.assert_called_once_with( - backup_update, update_mask, + backup_update, + update_mask, ) def test_update_expire_time_not_found(self): @@ -622,11 +637,13 @@ def test_update_expire_time_not_found(self): backup.update_expire_time(expire_time) backup_update = table_pb2.Backup( - name=self.BACKUP_NAME, expire_time=_datetime_to_pb_timestamp(expire_time), + name=self.BACKUP_NAME, + expire_time=_datetime_to_pb_timestamp(expire_time), ) update_mask = field_mask_pb2.FieldMask(paths=["expire_time"]) api.update_backup.assert_called_once_with( - backup_update, update_mask, + backup_update, + update_mask, ) def test_update_expire_time_success(self): @@ -644,11 +661,13 @@ def test_update_expire_time_success(self): backup.update_expire_time(expire_time) backup_update = table_pb2.Backup( - name=self.BACKUP_NAME, expire_time=_datetime_to_pb_timestamp(expire_time), + name=self.BACKUP_NAME, + expire_time=_datetime_to_pb_timestamp(expire_time), ) update_mask = field_mask_pb2.FieldMask(paths=["expire_time"]) api.update_backup.assert_called_once_with( - backup_update, update_mask, + backup_update, + update_mask, ) def test_restore_grpc_error(self): @@ -672,7 +691,9 @@ def test_restore_grpc_error(self): backup.restore(self.TABLE_ID) api.restore_table.assert_called_once_with( - parent=self.INSTANCE_NAME, table_id=self.TABLE_ID, backup=self.BACKUP_NAME, + parent=self.INSTANCE_NAME, + table_id=self.TABLE_ID, + backup=self.BACKUP_NAME, ) def test_restore_cluster_not_set(self): @@ -708,7 +729,9 @@ def test_restore_success(self): self.assertIs(future, op_future) api.restore_table.assert_called_once_with( - parent=self.INSTANCE_NAME, table_id=self.TABLE_ID, backup=self.BACKUP_NAME, + parent=self.INSTANCE_NAME, + table_id=self.TABLE_ID, + backup=self.BACKUP_NAME, ) diff --git a/tests/unit/test_policy.py b/tests/unit/test_policy.py index 63f9ba03f..939e02a9d 100644 --- a/tests/unit/test_policy.py +++ b/tests/unit/test_policy.py @@ -147,7 +147,11 @@ def test_from_pb_with_condition(self): }, } ] - message = policy_pb2.Policy(etag=ETAG, version=VERSION, bindings=BINDINGS,) + message = policy_pb2.Policy( + etag=ETAG, + version=VERSION, + bindings=BINDINGS, + ) klass = self._get_target_class() policy = klass.from_pb(message) self.assertEqual(policy.etag, ETAG) diff --git a/tests/unit/test_table.py b/tests/unit/test_table.py index f7377bc76..0ea45927d 100644 --- a/tests/unit/test_table.py +++ b/tests/unit/test_table.py @@ -1187,7 +1187,9 @@ def test_backup_factory_non_defaults(self): table = self._make_one(self.TABLE_ID, instance) timestamp = datetime.datetime.utcnow().replace(tzinfo=UTC) backup = table.backup( - self.BACKUP_ID, cluster_id=self.CLUSTER_ID, expire_time=timestamp, + self.BACKUP_ID, + cluster_id=self.CLUSTER_ID, + expire_time=timestamp, ) self.assertIsInstance(backup, Backup) @@ -1295,7 +1297,9 @@ def _restore_helper(self, backup_name=None): self.assertIs(future, op_future) api.restore_table.assert_called_once_with( - parent=self.INSTANCE_NAME, table_id=self.TABLE_ID, backup=self.BACKUP_NAME, + parent=self.INSTANCE_NAME, + table_id=self.TABLE_ID, + backup=self.BACKUP_NAME, ) def test_restore_table_w_backup_id(self):