From 4aa56ed75cfa92a4466fd54942242f98b56c8d87 Mon Sep 17 00:00:00 2001 From: Mend Renovate Date: Fri, 30 May 2025 23:55:47 +0200 Subject: [PATCH 01/10] chore(deps): update all dependencies (#947) * chore(deps): update all dependencies * Update docs.yml * Update lint.yml --------- Co-authored-by: Lingqing Gan --- samples/quickstart/requirements.txt | 2 +- samples/snippets/requirements.txt | 4 ++-- samples/to_dataframe/requirements.txt | 4 ++-- 3 files changed, 5 insertions(+), 5 deletions(-) diff --git a/samples/quickstart/requirements.txt b/samples/quickstart/requirements.txt index d3677224..7ddaaef7 100644 --- a/samples/quickstart/requirements.txt +++ b/samples/quickstart/requirements.txt @@ -1,2 +1,2 @@ fastavro -google-cloud-bigquery-storage==2.31.0 +google-cloud-bigquery-storage==2.32.0 diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt index 30eaa4f4..c34026ed 100644 --- a/samples/snippets/requirements.txt +++ b/samples/snippets/requirements.txt @@ -1,3 +1,3 @@ -google-cloud-bigquery-storage==2.31.0 +google-cloud-bigquery-storage==2.32.0 google-cloud-bigquery===3.30.0; python_version <= '3.8' -google-cloud-bigquery==3.33.0; python_version >= '3.9' +google-cloud-bigquery==3.34.0; python_version >= '3.9' diff --git a/samples/to_dataframe/requirements.txt b/samples/to_dataframe/requirements.txt index cc4d98c8..d12e3e9c 100644 --- a/samples/to_dataframe/requirements.txt +++ b/samples/to_dataframe/requirements.txt @@ -1,7 +1,7 @@ google-auth==2.40.2 -google-cloud-bigquery-storage==2.31.0 +google-cloud-bigquery-storage==2.32.0 google-cloud-bigquery===3.30.0; python_version <= '3.8' -google-cloud-bigquery==3.33.0; python_version >= '3.9' +google-cloud-bigquery==3.34.0; python_version >= '3.9' pyarrow===12.0.1; python_version == '3.7' pyarrow===17.0.0; python_version == '3.8' pyarrow==20.0.0; python_version >= '3.9' From 96a1743a998919ba652bd265689fa5c138d4286c Mon Sep 17 00:00:00 2001 From: Lingqing Gan Date: Fri, 1 Aug 2025 17:25:33 -0700 Subject: [PATCH 02/10] test: exclude lint.yml and docs.yml from renovate bot (#965) --- renovate.json | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/renovate.json b/renovate.json index c7875c46..9d9a6d0b 100644 --- a/renovate.json +++ b/renovate.json @@ -5,7 +5,7 @@ ":preserveSemverRanges", ":disableDependencyDashboard" ], - "ignorePaths": [".pre-commit-config.yaml", ".kokoro/requirements.txt", "setup.py", ".github/workflows/unittest.yml"], + "ignorePaths": [".pre-commit-config.yaml", ".kokoro/requirements.txt", "setup.py", ".github/workflows/unittest.yml", ".github/workflows/docs.yml", ".github/workflows/lint.yml"], "pip_requirements": { "fileMatch": ["requirements-test.txt", "samples/[\\S/]*constraints.txt", "samples/[\\S/]*constraints-test.txt"] } From 6e2c94fa863bd135f29407bee4ced491742b37ed Mon Sep 17 00:00:00 2001 From: Chalmer Lowe Date: Mon, 4 Aug 2025 18:20:43 -0400 Subject: [PATCH 03/10] feat: add support for Python 3.13 (#964) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * feat: add support for Python 3.13 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * updates errant owlbot/renovate.json issue that overwrites configs * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md * updates excludes for templated files * updates renovate.json * Update owlbot.py --------- Co-authored-by: google-labs-jules[bot] <161369871+google-labs-jules[bot]@users.noreply.github.com> Co-authored-by: Owl Bot --- .github/sync-repo-settings.yaml | 1 + owlbot.py | 2 +- setup.py | 1 + 3 files changed, 3 insertions(+), 1 deletion(-) diff --git a/.github/sync-repo-settings.yaml b/.github/sync-repo-settings.yaml index 439a0bcb..18d27d42 100644 --- a/.github/sync-repo-settings.yaml +++ b/.github/sync-repo-settings.yaml @@ -16,3 +16,4 @@ branchProtectionRules: - 'Samples - Python 3.10' - 'Samples - Python 3.11' - 'Samples - Python 3.12' + - 'Samples - Python 3.13' diff --git a/owlbot.py b/owlbot.py index be2f813a..646c9df3 100644 --- a/owlbot.py +++ b/owlbot.py @@ -139,7 +139,7 @@ ) s.move( templated_files, - excludes=[".coveragerc", ".github/release-please.yml", "docs/index.rst"], + excludes=[".coveragerc", ".github/release-please.yml", "docs/index.rst", "renovate.json",], ) python.py_samples(skip_readmes=True) diff --git a/setup.py b/setup.py index 88b5c5da..764c8b8b 100644 --- a/setup.py +++ b/setup.py @@ -86,6 +86,7 @@ "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", + "Programming Language :: Python :: 3.13", "Operating System :: OS Independent", "Topic :: Internet", ], From 132433ddb9a66dfe7ba893bb710686a06e98ee67 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Tim=20Swe=C3=B1a=20=28Swast=29?= Date: Thu, 14 Aug 2025 11:29:52 -0500 Subject: [PATCH 04/10] chore: test that ServiceUnavailable is retried (#970) * chore: test that ServiceUnavailable is retried * remove debug file --- ...st_client_v1.py => test_read_client_v1.py} | 56 +++++++++++++++---- 1 file changed, 45 insertions(+), 11 deletions(-) rename tests/unit/{test_client_v1.py => test_read_client_v1.py} (71%) diff --git a/tests/unit/test_client_v1.py b/tests/unit/test_read_client_v1.py similarity index 71% rename from tests/unit/test_client_v1.py rename to tests/unit/test_read_client_v1.py index f214b877..0c3e1dc1 100644 --- a/tests/unit/test_client_v1.py +++ b/tests/unit/test_read_client_v1.py @@ -15,11 +15,12 @@ import importlib from unittest import mock +import google.api_core.exceptions from google.api_core.gapic_v1 import client_info from google.auth import credentials import pytest -from google.cloud.bigquery_storage import types +from google.cloud.bigquery_storage_v1 import types PROJECT = "my-project" SERVICE_ACCOUNT_PROJECT = "project-from-credentials" @@ -29,20 +30,15 @@ def mock_transport(monkeypatch): from google.cloud.bigquery_storage_v1.services.big_query_read import transports - fake_create_session_rpc = mock.Mock(name="create_read_session_rpc") - fake_read_rows_rpc = mock.Mock(name="read_rows_rpc") - transport = mock.create_autospec( transports.grpc.BigQueryReadGrpcTransport, instance=True ) transport.create_read_session = mock.Mock(name="fake_create_read_session") transport.read_rows = mock.Mock(name="fake_read_rows") - - transport._wrapped_methods = { - transport.create_read_session: fake_create_session_rpc, - transport.read_rows: fake_read_rows_rpc, - } + transports.grpc.BigQueryReadGrpcTransport._prep_wrapped_messages( + transport, client_info.ClientInfo() + ) # _credentials property for TPC support transport._credentials = "" @@ -85,8 +81,11 @@ def __init__(self, *args, **kwargs): def test_create_read_session(mock_transport, client_under_test): - assert client_under_test._transport is mock_transport # sanity check + # validate test assumptions + assert client_under_test._transport is mock_transport + rpc_callable = mock.Mock() + mock_transport._wrapped_methods[mock_transport.create_read_session] = rpc_callable table = "projects/{}/datasets/{}/tables/{}".format( "data-project-id", "dataset_id", "table_id" ) @@ -101,12 +100,47 @@ def test_create_read_session(mock_transport, client_under_test): expected_session_arg = types.CreateReadSessionRequest( parent="projects/other-project", read_session=read_session ) - rpc_callable = mock_transport._wrapped_methods[mock_transport.create_read_session] rpc_callable.assert_called_once_with( expected_session_arg, metadata=mock.ANY, retry=mock.ANY, timeout=mock.ANY ) +def test_create_read_session_retries_serviceunavailable( + mock_transport, client_under_test +): + """Regression test for https://github.com/googleapis/python-bigquery-storage/issues/969.""" + # validate test assumptions + assert client_under_test._transport is mock_transport + + mock_transport.create_read_session.side_effect = [ + google.api_core.exceptions.ServiceUnavailable("connection reset"), + google.api_core.exceptions.ServiceUnavailable("connection reset"), + types.ReadSession(), + ] + table = "projects/{}/datasets/{}/tables/{}".format( + "data-project-id", "dataset_id", "table_id" + ) + read_session = types.ReadSession() + read_session.table = table + + # with pytest.raises(google.api_core.exceptions.ServiceUnavailable): + client_under_test.create_read_session( + parent="projects/other-project", read_session=read_session + ) + + expected_session_arg = types.CreateReadSessionRequest( + parent="projects/other-project", read_session=read_session + ) + expected_call = mock.call(expected_session_arg, metadata=mock.ANY, timeout=mock.ANY) + mock_transport.create_read_session.assert_has_calls( + [ + expected_call, + expected_call, + expected_call, + ] + ) + + def test_read_rows(mock_transport, client_under_test): stream_name = "teststream" offset = 0 From e2082ffc69f8753d858ac1c12b6e017fc3362fab Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Mon, 18 Aug 2025 10:32:18 -0700 Subject: [PATCH 05/10] feat: Add support for CMEK, runtime controls, and PSC-I to Reasoning Engine protos (#968) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * feat: Add support for CMEK, runtime controls, and PSC-I to Reasoning Engine protos docs: Add psc_interface_config, min/max_instances, resource_limits, container_concurrency to ReasoningEngineSpec docs: Add encryption_spec to ReasoningEngine PiperOrigin-RevId: 792745180 Source-Link: https://github.com/googleapis/googleapis/commit/85f7aecabe1da39d3377ab87135ea21c9169a6ea Source-Link: https://github.com/googleapis/googleapis-gen/commit/d68782f69bd428937f9bdc38cfc327e1457a85bf Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiZDY4NzgyZjY5YmQ0Mjg5MzdmOWJkYzM4Y2ZjMzI3ZTE0NTdhODViZiJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md --------- Co-authored-by: Owl Bot Co-authored-by: Lingqing Gan --- .../snippet_metadata_google.cloud.bigquery.storage.v1.json | 2 +- .../snippet_metadata_google.cloud.bigquery.storage.v1beta2.json | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1.json b/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1.json index 107928d6..c9e425d4 100644 --- a/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-bigquery-storage", - "version": "2.32.0" + "version": "0.1.0" }, "snippets": [ { diff --git a/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1beta2.json b/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1beta2.json index fce86775..e92a81e4 100644 --- a/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1beta2.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1beta2.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-bigquery-storage", - "version": "2.32.0" + "version": "0.1.0" }, "snippets": [ { From 7770c767d1a9918c1a1a2e5488d35c3de0a99e5e Mon Sep 17 00:00:00 2001 From: Mend Renovate Date: Mon, 18 Aug 2025 20:03:52 +0200 Subject: [PATCH 06/10] chore(deps): update all dependencies (#967) * chore(deps): update all dependencies * Update requirements-test.txt * Update requirements-test.txt * Update requirements-test.txt * Update requirements.txt * Update requirements-test.txt * Update requirements.txt --------- Co-authored-by: Lingqing Gan --- samples/pyarrow/requirements-test.txt | 3 ++- samples/quickstart/requirements-test.txt | 3 ++- samples/snippets/requirements-test.txt | 3 ++- samples/snippets/requirements.txt | 3 +++ samples/to_dataframe/requirements-test.txt | 3 ++- samples/to_dataframe/requirements.txt | 10 +++++----- 6 files changed, 16 insertions(+), 9 deletions(-) diff --git a/samples/pyarrow/requirements-test.txt b/samples/pyarrow/requirements-test.txt index 0404a8ad..7561ed55 100644 --- a/samples/pyarrow/requirements-test.txt +++ b/samples/pyarrow/requirements-test.txt @@ -1,2 +1,3 @@ pytest===7.4.3; python_version == '3.7' -pytest==8.3.5; python_version >= '3.8' +pytest===8.3.5; python_version == '3.8' +pytest==8.4.1; python_version >= '3.9' diff --git a/samples/quickstart/requirements-test.txt b/samples/quickstart/requirements-test.txt index 0404a8ad..7561ed55 100644 --- a/samples/quickstart/requirements-test.txt +++ b/samples/quickstart/requirements-test.txt @@ -1,2 +1,3 @@ pytest===7.4.3; python_version == '3.7' -pytest==8.3.5; python_version >= '3.8' +pytest===8.3.5; python_version == '3.8' +pytest==8.4.1; python_version >= '3.9' diff --git a/samples/snippets/requirements-test.txt b/samples/snippets/requirements-test.txt index a6c65ca6..230ca56d 100644 --- a/samples/snippets/requirements-test.txt +++ b/samples/snippets/requirements-test.txt @@ -1,3 +1,4 @@ google-cloud-testutils==1.6.4 pytest===7.4.3; python_version == '3.7' -pytest==8.3.5; python_version >= '3.8' +pytest===8.3.5; python_version == '3.8' +pytest==8.4.1; python_version >= '3.9' diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt index c34026ed..0ae6c4df 100644 --- a/samples/snippets/requirements.txt +++ b/samples/snippets/requirements.txt @@ -1,3 +1,6 @@ google-cloud-bigquery-storage==2.32.0 google-cloud-bigquery===3.30.0; python_version <= '3.8' google-cloud-bigquery==3.34.0; python_version >= '3.9' +pytest===7.4.3; python_version == '3.7' +pytest===8.3.5; python_version == '3.8' +pytest==8.4.1; python_version >= '3.9' diff --git a/samples/to_dataframe/requirements-test.txt b/samples/to_dataframe/requirements-test.txt index 0404a8ad..7561ed55 100644 --- a/samples/to_dataframe/requirements-test.txt +++ b/samples/to_dataframe/requirements-test.txt @@ -1,2 +1,3 @@ pytest===7.4.3; python_version == '3.7' -pytest==8.3.5; python_version >= '3.8' +pytest===8.3.5; python_version == '3.8' +pytest==8.4.1; python_version >= '3.9' diff --git a/samples/to_dataframe/requirements.txt b/samples/to_dataframe/requirements.txt index d12e3e9c..e3b75fda 100644 --- a/samples/to_dataframe/requirements.txt +++ b/samples/to_dataframe/requirements.txt @@ -1,19 +1,19 @@ -google-auth==2.40.2 +google-auth==2.40.3 google-cloud-bigquery-storage==2.32.0 google-cloud-bigquery===3.30.0; python_version <= '3.8' -google-cloud-bigquery==3.34.0; python_version >= '3.9' +google-cloud-bigquery==3.35.1; python_version >= '3.9' pyarrow===12.0.1; python_version == '3.7' pyarrow===17.0.0; python_version == '3.8' -pyarrow==20.0.0; python_version >= '3.9' +pyarrow==21.0.0; python_version >= '3.9' ipython===7.31.1; python_version == '3.7' ipython===8.10.0; python_version == '3.8' ipython===8.18.1; python_version == '3.9' ipython===8.33.0; python_version == '3.10' -ipython==9.2.0; python_version >= '3.11' +ipython==9.4.0; python_version >= '3.11' ipywidgets==8.1.7 pandas===1.3.5; python_version == '3.7' pandas===2.0.3; python_version == '3.8' -pandas==2.2.3; python_version >= '3.9' +pandas==2.3.1; python_version >= '3.9' tqdm==4.67.1 db-dtypes===1.4.2; python_version <= '3.8' db-dtypes==1.4.3; python_version >= '3.9' From 5d34e8cd4380f5ce3109ce2eaf3436942938c3a3 Mon Sep 17 00:00:00 2001 From: Mend Renovate Date: Mon, 18 Aug 2025 23:45:47 +0200 Subject: [PATCH 07/10] chore(deps): update dependency google-cloud-bigquery to v3.35.1 (#971) --- samples/snippets/requirements.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt index 0ae6c4df..8a456493 100644 --- a/samples/snippets/requirements.txt +++ b/samples/snippets/requirements.txt @@ -1,6 +1,6 @@ google-cloud-bigquery-storage==2.32.0 google-cloud-bigquery===3.30.0; python_version <= '3.8' -google-cloud-bigquery==3.34.0; python_version >= '3.9' +google-cloud-bigquery==3.35.1; python_version >= '3.9' pytest===7.4.3; python_version == '3.7' pytest===8.3.5; python_version == '3.8' pytest==8.4.1; python_version >= '3.9' From c106c73a8c10bf154ddf00e4a65ee2db2701aec1 Mon Sep 17 00:00:00 2001 From: Mend Renovate Date: Fri, 22 Aug 2025 00:17:44 +0200 Subject: [PATCH 08/10] chore(deps): update all dependencies (#972) --- samples/snippets/requirements.txt | 2 +- samples/to_dataframe/requirements.txt | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt index 8a456493..cab18ea2 100644 --- a/samples/snippets/requirements.txt +++ b/samples/snippets/requirements.txt @@ -1,6 +1,6 @@ google-cloud-bigquery-storage==2.32.0 google-cloud-bigquery===3.30.0; python_version <= '3.8' -google-cloud-bigquery==3.35.1; python_version >= '3.9' +google-cloud-bigquery==3.36.0; python_version >= '3.9' pytest===7.4.3; python_version == '3.7' pytest===8.3.5; python_version == '3.8' pytest==8.4.1; python_version >= '3.9' diff --git a/samples/to_dataframe/requirements.txt b/samples/to_dataframe/requirements.txt index e3b75fda..15e8e85c 100644 --- a/samples/to_dataframe/requirements.txt +++ b/samples/to_dataframe/requirements.txt @@ -1,7 +1,7 @@ google-auth==2.40.3 google-cloud-bigquery-storage==2.32.0 google-cloud-bigquery===3.30.0; python_version <= '3.8' -google-cloud-bigquery==3.35.1; python_version >= '3.9' +google-cloud-bigquery==3.36.0; python_version >= '3.9' pyarrow===12.0.1; python_version == '3.7' pyarrow===17.0.0; python_version == '3.8' pyarrow==21.0.0; python_version >= '3.9' @@ -13,7 +13,7 @@ ipython==9.4.0; python_version >= '3.11' ipywidgets==8.1.7 pandas===1.3.5; python_version == '3.7' pandas===2.0.3; python_version == '3.8' -pandas==2.3.1; python_version >= '3.9' +pandas==2.3.2; python_version >= '3.9' tqdm==4.67.1 db-dtypes===1.4.2; python_version <= '3.8' db-dtypes==1.4.3; python_version >= '3.9' From e2791c3b57922c70f6d54df9b6c8294f6fb6785a Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Tue, 2 Sep 2025 17:29:56 -0400 Subject: [PATCH 09/10] chore: Update gapic-generator-python to 1.26.2 (#976) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * chore: Update gapic-generator-python to 1.26.2 PiperOrigin-RevId: 802200836 Source-Link: https://github.com/googleapis/googleapis/commit/d300b151a973ce0425ae4ad07b3de957ca31bec6 Source-Link: https://github.com/googleapis/googleapis-gen/commit/a1ff0ae72ddcb68a259215d8c77661e2cdbb9b02 Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiYTFmZjBhZTcyZGRjYjY4YTI1OTIxNWQ4Yzc3NjYxZTJjZGJiOWIwMiJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md --------- Co-authored-by: Owl Bot --- .../services/big_query_write/async_client.py | 24 ++++++------- .../services/big_query_write/client.py | 24 ++++++------- .../big_query_write/transports/grpc.py | 24 ++++++------- .../transports/grpc_asyncio.py | 24 ++++++------- .../bigquery_storage_v1/types/storage.py | 34 +++++++++---------- .../cloud/bigquery_storage_v1/types/table.py | 30 ++++++++-------- .../types/metastore_partition.py | 10 +++--- .../services/big_query_write/async_client.py | 4 +-- .../services/big_query_write/client.py | 4 +-- .../big_query_write/transports/grpc.py | 4 +-- .../transports/grpc_asyncio.py | 4 +-- .../bigquery_storage_v1beta2/types/storage.py | 6 ++-- .../bigquery_storage_v1beta2/types/table.py | 2 +- 13 files changed, 97 insertions(+), 97 deletions(-) diff --git a/google/cloud/bigquery_storage_v1/services/big_query_write/async_client.py b/google/cloud/bigquery_storage_v1/services/big_query_write/async_client.py index fd526129..4d70a8cf 100644 --- a/google/cloud/bigquery_storage_v1/services/big_query_write/async_client.py +++ b/google/cloud/bigquery_storage_v1/services/big_query_write/async_client.py @@ -312,8 +312,8 @@ async def create_write_stream( metadata: Sequence[Tuple[str, Union[str, bytes]]] = (), ) -> stream.WriteStream: r"""Creates a write stream to the given table. Additionally, every - table has a special stream named '_default' to which data can be - written. This stream doesn't need to be created using + table has a special stream named '\_default' to which data can + be written. This stream doesn't need to be created using CreateWriteStream. It is a stream that can be used simultaneously by any number of clients. Data written to this stream is considered committed as soon as an acknowledgement is @@ -458,17 +458,17 @@ def append_rows( The specifics of when successfully appended data is made visible to the table are governed by the type of stream: - - For COMMITTED streams (which includes the default stream), - data is visible immediately upon successful append. + - For COMMITTED streams (which includes the default stream), + data is visible immediately upon successful append. - - For BUFFERED streams, data is made visible via a subsequent - ``FlushRows`` rpc which advances a cursor to a newer offset - in the stream. + - For BUFFERED streams, data is made visible via a subsequent + ``FlushRows`` rpc which advances a cursor to a newer offset in + the stream. - - For PENDING streams, data is not made visible until the - stream itself is finalized (via the ``FinalizeWriteStream`` - rpc), and the stream is explicitly committed via the - ``BatchCommitWriteStreams`` rpc. + - For PENDING streams, data is not made visible until the stream + itself is finalized (via the ``FinalizeWriteStream`` rpc), and + the stream is explicitly committed via the + ``BatchCommitWriteStreams`` rpc. .. code-block:: python @@ -678,7 +678,7 @@ async def finalize_write_stream( metadata: Sequence[Tuple[str, Union[str, bytes]]] = (), ) -> storage.FinalizeWriteStreamResponse: r"""Finalize a write stream so that no new data can be appended to - the stream. Finalize is not supported on the '_default' stream. + the stream. Finalize is not supported on the '\_default' stream. .. code-block:: python diff --git a/google/cloud/bigquery_storage_v1/services/big_query_write/client.py b/google/cloud/bigquery_storage_v1/services/big_query_write/client.py index 2ec4eb24..70cfad25 100644 --- a/google/cloud/bigquery_storage_v1/services/big_query_write/client.py +++ b/google/cloud/bigquery_storage_v1/services/big_query_write/client.py @@ -752,8 +752,8 @@ def create_write_stream( metadata: Sequence[Tuple[str, Union[str, bytes]]] = (), ) -> stream.WriteStream: r"""Creates a write stream to the given table. Additionally, every - table has a special stream named '_default' to which data can be - written. This stream doesn't need to be created using + table has a special stream named '\_default' to which data can + be written. This stream doesn't need to be created using CreateWriteStream. It is a stream that can be used simultaneously by any number of clients. Data written to this stream is considered committed as soon as an acknowledgement is @@ -895,17 +895,17 @@ def append_rows( The specifics of when successfully appended data is made visible to the table are governed by the type of stream: - - For COMMITTED streams (which includes the default stream), - data is visible immediately upon successful append. + - For COMMITTED streams (which includes the default stream), + data is visible immediately upon successful append. - - For BUFFERED streams, data is made visible via a subsequent - ``FlushRows`` rpc which advances a cursor to a newer offset - in the stream. + - For BUFFERED streams, data is made visible via a subsequent + ``FlushRows`` rpc which advances a cursor to a newer offset in + the stream. - - For PENDING streams, data is not made visible until the - stream itself is finalized (via the ``FinalizeWriteStream`` - rpc), and the stream is explicitly committed via the - ``BatchCommitWriteStreams`` rpc. + - For PENDING streams, data is not made visible until the stream + itself is finalized (via the ``FinalizeWriteStream`` rpc), and + the stream is explicitly committed via the + ``BatchCommitWriteStreams`` rpc. .. code-block:: python @@ -1110,7 +1110,7 @@ def finalize_write_stream( metadata: Sequence[Tuple[str, Union[str, bytes]]] = (), ) -> storage.FinalizeWriteStreamResponse: r"""Finalize a write stream so that no new data can be appended to - the stream. Finalize is not supported on the '_default' stream. + the stream. Finalize is not supported on the '\_default' stream. .. code-block:: python diff --git a/google/cloud/bigquery_storage_v1/services/big_query_write/transports/grpc.py b/google/cloud/bigquery_storage_v1/services/big_query_write/transports/grpc.py index c02e6d47..7ab45f9f 100644 --- a/google/cloud/bigquery_storage_v1/services/big_query_write/transports/grpc.py +++ b/google/cloud/bigquery_storage_v1/services/big_query_write/transports/grpc.py @@ -332,8 +332,8 @@ def create_write_stream( r"""Return a callable for the create write stream method over gRPC. Creates a write stream to the given table. Additionally, every - table has a special stream named '_default' to which data can be - written. This stream doesn't need to be created using + table has a special stream named '\_default' to which data can + be written. This stream doesn't need to be created using CreateWriteStream. It is a stream that can be used simultaneously by any number of clients. Data written to this stream is considered committed as soon as an acknowledgement is @@ -386,17 +386,17 @@ def append_rows( The specifics of when successfully appended data is made visible to the table are governed by the type of stream: - - For COMMITTED streams (which includes the default stream), - data is visible immediately upon successful append. + - For COMMITTED streams (which includes the default stream), + data is visible immediately upon successful append. - - For BUFFERED streams, data is made visible via a subsequent - ``FlushRows`` rpc which advances a cursor to a newer offset - in the stream. + - For BUFFERED streams, data is made visible via a subsequent + ``FlushRows`` rpc which advances a cursor to a newer offset in + the stream. - - For PENDING streams, data is not made visible until the - stream itself is finalized (via the ``FinalizeWriteStream`` - rpc), and the stream is explicitly committed via the - ``BatchCommitWriteStreams`` rpc. + - For PENDING streams, data is not made visible until the stream + itself is finalized (via the ``FinalizeWriteStream`` rpc), and + the stream is explicitly committed via the + ``BatchCommitWriteStreams`` rpc. Returns: Callable[[~.AppendRowsRequest], @@ -451,7 +451,7 @@ def finalize_write_stream( r"""Return a callable for the finalize write stream method over gRPC. Finalize a write stream so that no new data can be appended to - the stream. Finalize is not supported on the '_default' stream. + the stream. Finalize is not supported on the '\_default' stream. Returns: Callable[[~.FinalizeWriteStreamRequest], diff --git a/google/cloud/bigquery_storage_v1/services/big_query_write/transports/grpc_asyncio.py b/google/cloud/bigquery_storage_v1/services/big_query_write/transports/grpc_asyncio.py index b066bb55..1337dcdd 100644 --- a/google/cloud/bigquery_storage_v1/services/big_query_write/transports/grpc_asyncio.py +++ b/google/cloud/bigquery_storage_v1/services/big_query_write/transports/grpc_asyncio.py @@ -340,8 +340,8 @@ def create_write_stream( r"""Return a callable for the create write stream method over gRPC. Creates a write stream to the given table. Additionally, every - table has a special stream named '_default' to which data can be - written. This stream doesn't need to be created using + table has a special stream named '\_default' to which data can + be written. This stream doesn't need to be created using CreateWriteStream. It is a stream that can be used simultaneously by any number of clients. Data written to this stream is considered committed as soon as an acknowledgement is @@ -394,17 +394,17 @@ def append_rows( The specifics of when successfully appended data is made visible to the table are governed by the type of stream: - - For COMMITTED streams (which includes the default stream), - data is visible immediately upon successful append. + - For COMMITTED streams (which includes the default stream), + data is visible immediately upon successful append. - - For BUFFERED streams, data is made visible via a subsequent - ``FlushRows`` rpc which advances a cursor to a newer offset - in the stream. + - For BUFFERED streams, data is made visible via a subsequent + ``FlushRows`` rpc which advances a cursor to a newer offset in + the stream. - - For PENDING streams, data is not made visible until the - stream itself is finalized (via the ``FinalizeWriteStream`` - rpc), and the stream is explicitly committed via the - ``BatchCommitWriteStreams`` rpc. + - For PENDING streams, data is not made visible until the stream + itself is finalized (via the ``FinalizeWriteStream`` rpc), and + the stream is explicitly committed via the + ``BatchCommitWriteStreams`` rpc. Returns: Callable[[~.AppendRowsRequest], @@ -460,7 +460,7 @@ def finalize_write_stream( r"""Return a callable for the finalize write stream method over gRPC. Finalize a write stream so that no new data can be appended to - the stream. Finalize is not supported on the '_default' stream. + the stream. Finalize is not supported on the '\_default' stream. Returns: Callable[[~.FinalizeWriteStreamRequest], diff --git a/google/cloud/bigquery_storage_v1/types/storage.py b/google/cloud/bigquery_storage_v1/types/storage.py index ee0e4e14..da966f53 100644 --- a/google/cloud/bigquery_storage_v1/types/storage.py +++ b/google/cloud/bigquery_storage_v1/types/storage.py @@ -333,11 +333,11 @@ class SplitReadStreamResponse(proto.Message): Attributes: primary_stream (google.cloud.bigquery_storage_v1.types.ReadStream): Primary stream, which contains the beginning portion of - \|original_stream|. An empty value indicates that the + \|original_stream\|. An empty value indicates that the original stream can no longer be split. remainder_stream (google.cloud.bigquery_storage_v1.types.ReadStream): Remainder stream, which contains the tail of - \|original_stream|. An empty value indicates that the + \|original_stream\|. An empty value indicates that the original stream can no longer be split. """ @@ -401,32 +401,32 @@ class AppendRowsRequest(proto.Message): Required. The write_stream identifies the append operation. It must be provided in the following scenarios: - - In the first request to an AppendRows connection. + - In the first request to an AppendRows connection. - - In all subsequent requests to an AppendRows connection, - if you use the same connection to write to multiple - tables or change the input schema for default streams. + - In all subsequent requests to an AppendRows connection, if + you use the same connection to write to multiple tables or + change the input schema for default streams. For explicitly created write streams, the format is: - - ``projects/{project}/datasets/{dataset}/tables/{table}/streams/{id}`` + - ``projects/{project}/datasets/{dataset}/tables/{table}/streams/{id}`` For the special default stream, the format is: - - ``projects/{project}/datasets/{dataset}/tables/{table}/streams/_default``. + - ``projects/{project}/datasets/{dataset}/tables/{table}/streams/_default``. An example of a possible sequence of requests with write_stream fields within a single connection: - - r1: {write_stream: stream_name_1} + - r1: {write_stream: stream_name_1} - - r2: {write_stream: /*omit*/} + - r2: {write_stream: /*omit*/} - - r3: {write_stream: /*omit*/} + - r3: {write_stream: /*omit*/} - - r4: {write_stream: stream_name_2} + - r4: {write_stream: stream_name_2} - - r5: {write_stream: stream_name_2} + - r5: {write_stream: stream_name_2} The destination changed in request_4, so the write_stream field must be populated in all subsequent requests in this @@ -436,7 +436,7 @@ class AppendRowsRequest(proto.Message): offset is same as the provided value. If not present, the write is performed at the current end of stream. Specifying a value for this field is not allowed when calling - AppendRows for the '_default' stream. + AppendRows for the '\_default' stream. proto_rows (google.cloud.bigquery_storage_v1.types.AppendRowsRequest.ProtoData): Rows in proto format. @@ -539,11 +539,11 @@ class ProtoData(proto.Message): The protocol buffer schema used to serialize the data. Provide this value whenever: - - You send the first request of an RPC connection. + - You send the first request of an RPC connection. - - You change the input schema. + - You change the input schema. - - You specify a new destination table. + - You specify a new destination table. rows (google.cloud.bigquery_storage_v1.types.ProtoRows): Serialized row data in protobuf message format. Currently, the backend expects the diff --git a/google/cloud/bigquery_storage_v1/types/table.py b/google/cloud/bigquery_storage_v1/types/table.py index bce1a6ad..a902b7d2 100644 --- a/google/cloud/bigquery_storage_v1/types/table.py +++ b/google/cloud/bigquery_storage_v1/types/table.py @@ -52,7 +52,7 @@ class TableFieldSchema(proto.Message): Attributes: name (str): Required. The field name. The name must contain only letters - (a-z, A-Z), numbers (0-9), or underscores (_), and must + (a-z, A-Z), numbers (0-9), or underscores (\_), and must start with a letter or underscore. The maximum length is 128 characters. type_ (google.cloud.bigquery_storage_v1.types.TableFieldSchema.Type): @@ -97,26 +97,26 @@ class TableFieldSchema(proto.Message): Values of this NUMERIC or BIGNUMERIC field must be in this range when: - - Precision (P) and scale (S) are specified: [-10^(P-S) + - 10^(-S), 10^(P-S) - 10^(-S)] - - Precision (P) is specified but not scale (and thus scale - is interpreted to be equal to zero): [-10^P + 1, 10^P - - 1]. + - Precision (P) and scale (S) are specified: [-10^(P-S) + + 10^(-S), 10^(P-S) - 10^(-S)] + - Precision (P) is specified but not scale (and thus scale + is interpreted to be equal to zero): [-10^P + 1, 10^P - + 1]. Acceptable values for precision and scale if both are specified: - - If type = "NUMERIC": 1 <= precision - scale <= 29 and 0 - <= scale <= 9. - - If type = "BIGNUMERIC": 1 <= precision - scale <= 38 and - 0 <= scale <= 38. + - If type = "NUMERIC": 1 <= precision - scale <= 29 and 0 <= + scale <= 9. + - If type = "BIGNUMERIC": 1 <= precision - scale <= 38 and 0 + <= scale <= 38. Acceptable values for precision if only precision is specified but not scale (and thus scale is interpreted to be equal to zero): - - If type = "NUMERIC": 1 <= precision <= 29. - - If type = "BIGNUMERIC": 1 <= precision <= 38. + - If type = "NUMERIC": 1 <= precision <= 29. + - If type = "BIGNUMERIC": 1 <= precision <= 38. If scale is specified but not precision, then it is invalid. scale (int): @@ -131,9 +131,9 @@ class TableFieldSchema(proto.Message): required. Possible values for the field element type of a RANGE include: - - DATE - - DATETIME - - TIMESTAMP + - DATE + - DATETIME + - TIMESTAMP """ class Type(proto.Enum): diff --git a/google/cloud/bigquery_storage_v1beta/types/metastore_partition.py b/google/cloud/bigquery_storage_v1beta/types/metastore_partition.py index 14ecc5e5..bae4198b 100644 --- a/google/cloud/bigquery_storage_v1beta/types/metastore_partition.py +++ b/google/cloud/bigquery_storage_v1beta/types/metastore_partition.py @@ -267,11 +267,11 @@ class ListMetastorePartitionsRequest(proto.Message): Examples: - - "int_field > 5" - - "date_field = CAST('2014-9-27' as DATE)" - - "nullable_field is not NULL" - - "st_equals(geo_field, st_geofromtext("POINT(2, 2)"))" - - "numeric_field BETWEEN 1.0 AND 5.0" + - "int_field > 5" + - "date_field = CAST('2014-9-27' as DATE)" + - "nullable_field is not NULL" + - "st_equals(geo_field, st_geofromtext("POINT(2, 2)"))" + - "numeric_field BETWEEN 1.0 AND 5.0" Restricted to a maximum length of 1 MB. trace_id (str): diff --git a/google/cloud/bigquery_storage_v1beta2/services/big_query_write/async_client.py b/google/cloud/bigquery_storage_v1beta2/services/big_query_write/async_client.py index 3469dbae..6f48fd18 100644 --- a/google/cloud/bigquery_storage_v1beta2/services/big_query_write/async_client.py +++ b/google/cloud/bigquery_storage_v1beta2/services/big_query_write/async_client.py @@ -314,7 +314,7 @@ async def create_write_stream( metadata: Sequence[Tuple[str, Union[str, bytes]]] = (), ) -> stream.WriteStream: r"""Creates a write stream to the given table. Additionally, every - table has a special COMMITTED stream named '_default' to which + table has a special COMMITTED stream named '\_default' to which data can be written. This stream doesn't need to be created using CreateWriteStream. It is a stream that can be used simultaneously by any number of clients. Data written to this @@ -670,7 +670,7 @@ async def finalize_write_stream( metadata: Sequence[Tuple[str, Union[str, bytes]]] = (), ) -> storage.FinalizeWriteStreamResponse: r"""Finalize a write stream so that no new data can be appended to - the stream. Finalize is not supported on the '_default' stream. + the stream. Finalize is not supported on the '\_default' stream. .. code-block:: python diff --git a/google/cloud/bigquery_storage_v1beta2/services/big_query_write/client.py b/google/cloud/bigquery_storage_v1beta2/services/big_query_write/client.py index 849e44e3..4603aa55 100644 --- a/google/cloud/bigquery_storage_v1beta2/services/big_query_write/client.py +++ b/google/cloud/bigquery_storage_v1beta2/services/big_query_write/client.py @@ -753,7 +753,7 @@ def create_write_stream( metadata: Sequence[Tuple[str, Union[str, bytes]]] = (), ) -> stream.WriteStream: r"""Creates a write stream to the given table. Additionally, every - table has a special COMMITTED stream named '_default' to which + table has a special COMMITTED stream named '\_default' to which data can be written. This stream doesn't need to be created using CreateWriteStream. It is a stream that can be used simultaneously by any number of clients. Data written to this @@ -1099,7 +1099,7 @@ def finalize_write_stream( metadata: Sequence[Tuple[str, Union[str, bytes]]] = (), ) -> storage.FinalizeWriteStreamResponse: r"""Finalize a write stream so that no new data can be appended to - the stream. Finalize is not supported on the '_default' stream. + the stream. Finalize is not supported on the '\_default' stream. .. code-block:: python diff --git a/google/cloud/bigquery_storage_v1beta2/services/big_query_write/transports/grpc.py b/google/cloud/bigquery_storage_v1beta2/services/big_query_write/transports/grpc.py index 33ab7b39..2abe904b 100644 --- a/google/cloud/bigquery_storage_v1beta2/services/big_query_write/transports/grpc.py +++ b/google/cloud/bigquery_storage_v1beta2/services/big_query_write/transports/grpc.py @@ -333,7 +333,7 @@ def create_write_stream( r"""Return a callable for the create write stream method over gRPC. Creates a write stream to the given table. Additionally, every - table has a special COMMITTED stream named '_default' to which + table has a special COMMITTED stream named '\_default' to which data can be written. This stream doesn't need to be created using CreateWriteStream. It is a stream that can be used simultaneously by any number of clients. Data written to this @@ -439,7 +439,7 @@ def finalize_write_stream( r"""Return a callable for the finalize write stream method over gRPC. Finalize a write stream so that no new data can be appended to - the stream. Finalize is not supported on the '_default' stream. + the stream. Finalize is not supported on the '\_default' stream. Returns: Callable[[~.FinalizeWriteStreamRequest], diff --git a/google/cloud/bigquery_storage_v1beta2/services/big_query_write/transports/grpc_asyncio.py b/google/cloud/bigquery_storage_v1beta2/services/big_query_write/transports/grpc_asyncio.py index 65593db6..5f3b017b 100644 --- a/google/cloud/bigquery_storage_v1beta2/services/big_query_write/transports/grpc_asyncio.py +++ b/google/cloud/bigquery_storage_v1beta2/services/big_query_write/transports/grpc_asyncio.py @@ -341,7 +341,7 @@ def create_write_stream( r"""Return a callable for the create write stream method over gRPC. Creates a write stream to the given table. Additionally, every - table has a special COMMITTED stream named '_default' to which + table has a special COMMITTED stream named '\_default' to which data can be written. This stream doesn't need to be created using CreateWriteStream. It is a stream that can be used simultaneously by any number of clients. Data written to this @@ -448,7 +448,7 @@ def finalize_write_stream( r"""Return a callable for the finalize write stream method over gRPC. Finalize a write stream so that no new data can be appended to - the stream. Finalize is not supported on the '_default' stream. + the stream. Finalize is not supported on the '\_default' stream. Returns: Callable[[~.FinalizeWriteStreamRequest], diff --git a/google/cloud/bigquery_storage_v1beta2/types/storage.py b/google/cloud/bigquery_storage_v1beta2/types/storage.py index 87aa4218..11145e5e 100644 --- a/google/cloud/bigquery_storage_v1beta2/types/storage.py +++ b/google/cloud/bigquery_storage_v1beta2/types/storage.py @@ -295,11 +295,11 @@ class SplitReadStreamResponse(proto.Message): Attributes: primary_stream (google.cloud.bigquery_storage_v1beta2.types.ReadStream): Primary stream, which contains the beginning portion of - \|original_stream|. An empty value indicates that the + \|original_stream\|. An empty value indicates that the original stream can no longer be split. remainder_stream (google.cloud.bigquery_storage_v1beta2.types.ReadStream): Remainder stream, which contains the tail of - \|original_stream|. An empty value indicates that the + \|original_stream\|. An empty value indicates that the original stream can no longer be split. """ @@ -357,7 +357,7 @@ class AppendRowsRequest(proto.Message): offset is same as the provided value. If not present, the write is performed at the current end of stream. Specifying a value for this field is not allowed when calling - AppendRows for the '_default' stream. + AppendRows for the '\_default' stream. proto_rows (google.cloud.bigquery_storage_v1beta2.types.AppendRowsRequest.ProtoData): Rows in proto format. diff --git a/google/cloud/bigquery_storage_v1beta2/types/table.py b/google/cloud/bigquery_storage_v1beta2/types/table.py index 43c7d755..4263b654 100644 --- a/google/cloud/bigquery_storage_v1beta2/types/table.py +++ b/google/cloud/bigquery_storage_v1beta2/types/table.py @@ -49,7 +49,7 @@ class TableFieldSchema(proto.Message): Attributes: name (str): Required. The field name. The name must contain only letters - (a-z, A-Z), numbers (0-9), or underscores (_), and must + (a-z, A-Z), numbers (0-9), or underscores (\_), and must start with a letter or underscore. The maximum length is 128 characters. type_ (google.cloud.bigquery_storage_v1beta2.types.TableFieldSchema.Type): From 6c8408e56e4b8a598666e7b7b264fd6afa745403 Mon Sep 17 00:00:00 2001 From: "release-please[bot]" <55107282+release-please[bot]@users.noreply.github.com> Date: Wed, 3 Sep 2025 12:40:46 -0400 Subject: [PATCH 10/10] chore(main): release 2.33.0 (#966) Co-authored-by: release-please[bot] <55107282+release-please[bot]@users.noreply.github.com> --- .release-please-manifest.json | 2 +- CHANGELOG.md | 8 ++++++++ google/cloud/bigquery_storage/gapic_version.py | 2 +- google/cloud/bigquery_storage_v1/gapic_version.py | 2 +- google/cloud/bigquery_storage_v1alpha/gapic_version.py | 2 +- google/cloud/bigquery_storage_v1beta/gapic_version.py | 2 +- google/cloud/bigquery_storage_v1beta2/gapic_version.py | 2 +- ...snippet_metadata_google.cloud.bigquery.storage.v1.json | 2 +- ...et_metadata_google.cloud.bigquery.storage.v1beta2.json | 2 +- 9 files changed, 16 insertions(+), 8 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 7bc1375d..7d35f2a0 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,4 +1,4 @@ { - ".": "2.32.0" + ".": "2.33.0" } \ No newline at end of file diff --git a/CHANGELOG.md b/CHANGELOG.md index 64239470..f81bef13 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,6 +4,14 @@ [1]: https://pypi.org/project/google-cloud-bigquery-storage/#history +## [2.33.0](https://github.com/googleapis/python-bigquery-storage/compare/v2.32.0...v2.33.0) (2025-09-02) + + +### Features + +* Add support for CMEK, runtime controls, and PSC-I to Reasoning Engine protos ([#968](https://github.com/googleapis/python-bigquery-storage/issues/968)) ([e2082ff](https://github.com/googleapis/python-bigquery-storage/commit/e2082ffc69f8753d858ac1c12b6e017fc3362fab)) +* Add support for Python 3.13 ([#964](https://github.com/googleapis/python-bigquery-storage/issues/964)) ([6e2c94f](https://github.com/googleapis/python-bigquery-storage/commit/6e2c94fa863bd135f29407bee4ced491742b37ed)) + ## [2.32.0](https://github.com/googleapis/python-bigquery-storage/compare/v2.31.0...v2.32.0) (2025-05-27) diff --git a/google/cloud/bigquery_storage/gapic_version.py b/google/cloud/bigquery_storage/gapic_version.py index 3c958586..0c5de5c0 100644 --- a/google/cloud/bigquery_storage/gapic_version.py +++ b/google/cloud/bigquery_storage/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "2.32.0" # {x-release-please-version} +__version__ = "2.33.0" # {x-release-please-version} diff --git a/google/cloud/bigquery_storage_v1/gapic_version.py b/google/cloud/bigquery_storage_v1/gapic_version.py index 3c958586..0c5de5c0 100644 --- a/google/cloud/bigquery_storage_v1/gapic_version.py +++ b/google/cloud/bigquery_storage_v1/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "2.32.0" # {x-release-please-version} +__version__ = "2.33.0" # {x-release-please-version} diff --git a/google/cloud/bigquery_storage_v1alpha/gapic_version.py b/google/cloud/bigquery_storage_v1alpha/gapic_version.py index 3c958586..0c5de5c0 100644 --- a/google/cloud/bigquery_storage_v1alpha/gapic_version.py +++ b/google/cloud/bigquery_storage_v1alpha/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "2.32.0" # {x-release-please-version} +__version__ = "2.33.0" # {x-release-please-version} diff --git a/google/cloud/bigquery_storage_v1beta/gapic_version.py b/google/cloud/bigquery_storage_v1beta/gapic_version.py index 2e49dfda..1eb64a9d 100644 --- a/google/cloud/bigquery_storage_v1beta/gapic_version.py +++ b/google/cloud/bigquery_storage_v1beta/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "2.32.0" # {x-release-please-version} +__version__ = "2.33.0" # {x-release-please-version} diff --git a/google/cloud/bigquery_storage_v1beta2/gapic_version.py b/google/cloud/bigquery_storage_v1beta2/gapic_version.py index 3c958586..0c5de5c0 100644 --- a/google/cloud/bigquery_storage_v1beta2/gapic_version.py +++ b/google/cloud/bigquery_storage_v1beta2/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "2.32.0" # {x-release-please-version} +__version__ = "2.33.0" # {x-release-please-version} diff --git a/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1.json b/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1.json index c9e425d4..e7eaf5fe 100644 --- a/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-bigquery-storage", - "version": "0.1.0" + "version": "2.33.0" }, "snippets": [ { diff --git a/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1beta2.json b/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1beta2.json index e92a81e4..db1af5cb 100644 --- a/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1beta2.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.bigquery.storage.v1beta2.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-bigquery-storage", - "version": "0.1.0" + "version": "2.33.0" }, "snippets": [ {