Skip to content
This repository was archived by the owner on Mar 31, 2026. It is now read-only.
Merged
7 changes: 5 additions & 2 deletions samples/samples/autocommit.py
Original file line number Diff line number Diff line change
Expand Up @@ -46,11 +46,14 @@ def enable_autocommit_mode(instance_id, database_id):

if __name__ == "__main__":
parser = argparse.ArgumentParser(
description=__doc__, formatter_class=argparse.RawDescriptionHelpFormatter,
description=__doc__,
formatter_class=argparse.RawDescriptionHelpFormatter,
)
parser.add_argument("instance_id", help="Your Cloud Spanner instance ID.")
parser.add_argument(
"--database-id", help="Your Cloud Spanner database ID.", default="example_db",
"--database-id",
help="Your Cloud Spanner database ID.",
default="example_db",
)
subparsers = parser.add_subparsers(dest="command")
subparsers.add_parser("enable_autocommit_mode", help=enable_autocommit_mode.__doc__)
Expand Down
3 changes: 2 additions & 1 deletion samples/samples/autocommit_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,8 @@ def test_enable_autocommit_mode(capsys, instance_id, sample_database):
op.result()

autocommit.enable_autocommit_mode(
instance_id, sample_database.database_id,
instance_id,
sample_database.database_id,
)
out, _ = capsys.readouterr()
assert "Autocommit mode is enabled." in out
Expand Down
40 changes: 30 additions & 10 deletions samples/samples/backup_sample_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -26,12 +26,12 @@ def sample_name():


def unique_database_id():
""" Creates a unique id for the database. """
"""Creates a unique id for the database."""
return f"test-db-{uuid.uuid4().hex[:10]}"


def unique_backup_id():
""" Creates a unique id for the backup. """
"""Creates a unique id for the backup."""
return f"test-backup-{uuid.uuid4().hex[:10]}"


Expand All @@ -52,7 +52,10 @@ def test_create_backup(capsys, instance_id, sample_database):
version_time = list(results)[0][0]

backup_sample.create_backup(
instance_id, sample_database.database_id, BACKUP_ID, version_time,
instance_id,
sample_database.database_id,
BACKUP_ID,
version_time,
)
out, _ = capsys.readouterr()
assert BACKUP_ID in out
Expand All @@ -74,10 +77,16 @@ def test_copy_backup(capsys, instance_id, spanner_client):

@pytest.mark.dependency(name="create_backup_with_encryption_key")
def test_create_backup_with_encryption_key(
capsys, instance_id, sample_database, kms_key_name,
capsys,
instance_id,
sample_database,
kms_key_name,
):
backup_sample.create_backup_with_encryption_key(
instance_id, sample_database.database_id, CMEK_BACKUP_ID, kms_key_name,
instance_id,
sample_database.database_id,
CMEK_BACKUP_ID,
kms_key_name,
)
out, _ = capsys.readouterr()
assert CMEK_BACKUP_ID in out
Expand All @@ -97,7 +106,10 @@ def test_restore_database(capsys, instance_id, sample_database):
@pytest.mark.dependency(depends=["create_backup_with_encryption_key"])
@RetryErrors(exception=DeadlineExceeded, max_tries=2)
def test_restore_database_with_encryption_key(
capsys, instance_id, sample_database, kms_key_name,
capsys,
instance_id,
sample_database,
kms_key_name,
):
backup_sample.restore_database_with_encryption_key(
instance_id, CMEK_RESTORE_DB_ID, CMEK_BACKUP_ID, kms_key_name
Expand All @@ -123,10 +135,14 @@ def test_list_backup_operations(capsys, instance_id, sample_database):

@pytest.mark.dependency(name="list_backup", depends=["create_backup", "copy_backup"])
def test_list_backups(
capsys, instance_id, sample_database,
capsys,
instance_id,
sample_database,
):
backup_sample.list_backups(
instance_id, sample_database.database_id, BACKUP_ID,
instance_id,
sample_database.database_id,
BACKUP_ID,
)
out, _ = capsys.readouterr()
id_count = out.count(BACKUP_ID)
Expand All @@ -153,7 +169,9 @@ def test_delete_backup(capsys, instance_id):
@pytest.mark.dependency(depends=["create_backup"])
def test_cancel_backup(capsys, instance_id, sample_database):
backup_sample.cancel_backup(
instance_id, sample_database.database_id, BACKUP_ID,
instance_id,
sample_database.database_id,
BACKUP_ID,
)
out, _ = capsys.readouterr()
cancel_success = "Backup creation was successfully cancelled." in out
Expand All @@ -166,7 +184,9 @@ def test_cancel_backup(capsys, instance_id, sample_database):
@RetryErrors(exception=DeadlineExceeded, max_tries=2)
def test_create_database_with_retention_period(capsys, sample_instance):
backup_sample.create_database_with_version_retention_period(
sample_instance.instance_id, RETENTION_DATABASE_ID, RETENTION_PERIOD,
sample_instance.instance_id,
RETENTION_DATABASE_ID,
RETENTION_PERIOD,
)
out, _ = capsys.readouterr()
assert (RETENTION_DATABASE_ID + " created with ") in out
Expand Down
4 changes: 2 additions & 2 deletions samples/samples/batch_sample.py
Original file line number Diff line number Diff line change
Expand Up @@ -57,7 +57,7 @@ def run_batch_query(instance_id, database_id):
for future in concurrent.futures.as_completed(futures, timeout=3600):
finish, row_ct = future.result()
elapsed = finish - start
print(u"Completed {} rows in {} seconds".format(row_ct, elapsed))
print("Completed {} rows in {} seconds".format(row_ct, elapsed))

# Clean up
snapshot.close()
Expand All @@ -68,7 +68,7 @@ def process(snapshot, partition):
print("Started processing partition.")
row_ct = 0
for row in snapshot.process_read_batch(partition):
print(u"SingerId: {}, AlbumId: {}, AlbumTitle: {}".format(*row))
print("SingerId: {}, AlbumId: {}, AlbumTitle: {}".format(*row))
row_ct += 1
return time.time(), row_ct

Expand Down
11 changes: 8 additions & 3 deletions samples/samples/conftest.py
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@

@pytest.fixture(scope="module")
def sample_name():
""" Sample testcase modules must define this fixture.
"""Sample testcase modules must define this fixture.

The name is used to label the instance created by the sample, to
aid in debugging leaked instances.
Expand Down Expand Up @@ -98,7 +98,11 @@ def multi_region_instance_config(spanner_client):

@pytest.fixture(scope="module")
def sample_instance(
spanner_client, cleanup_old_instances, instance_id, instance_config, sample_name,
spanner_client,
cleanup_old_instances,
instance_id,
instance_config,
sample_name,
):
sample_instance = spanner_client.instance(
instance_id,
Expand Down Expand Up @@ -184,7 +188,8 @@ def database_ddl():
def sample_database(sample_instance, database_id, database_ddl):

sample_database = sample_instance.database(
database_id, ddl_statements=database_ddl,
database_id,
ddl_statements=database_ddl,
)

if not sample_database.exists():
Expand Down
Loading