diff --git a/.flake8 b/.flake8 index 29227d4c..2e438749 100644 --- a/.flake8 +++ b/.flake8 @@ -16,7 +16,7 @@ # Generated by synthtool. DO NOT EDIT! [flake8] -ignore = E203, E266, E501, W503 +ignore = E203, E231, E266, E501, W503 exclude = # Exclude generated code. **/proto/** diff --git a/.github/.OwlBot.lock.yaml b/.github/.OwlBot.lock.yaml index 44c78f7c..757c9dca 100644 --- a/.github/.OwlBot.lock.yaml +++ b/.github/.OwlBot.lock.yaml @@ -13,4 +13,5 @@ # limitations under the License. docker: image: gcr.io/cloud-devrel-public-resources/owlbot-python:latest - digest: sha256:4e1991042fe54b991db9ca17c8fb386e61b22fe4d1472a568bf0fcac85dcf5d3 + digest: sha256:81ed5ecdfc7cac5b699ba4537376f3563f6f04122c4ec9e735d3b3dc1d43dd32 +# created: 2022-05-05T22:08:23.383410683Z diff --git a/.github/auto-approve.yml b/.github/auto-approve.yml new file mode 100644 index 00000000..311ebbb8 --- /dev/null +++ b/.github/auto-approve.yml @@ -0,0 +1,3 @@ +# https://github.com/googleapis/repo-automation-bots/tree/main/packages/auto-approve +processes: + - "OwlBotTemplateChanges" diff --git a/.github/auto-label.yaml b/.github/auto-label.yaml new file mode 100644 index 00000000..41bff0b5 --- /dev/null +++ b/.github/auto-label.yaml @@ -0,0 +1,15 @@ +# Copyright 2022 Google LLC +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +requestsize: + enabled: true diff --git a/.github/release-please.yml b/.github/release-please.yml index 466597e5..5161ab34 100644 --- a/.github/release-please.yml +++ b/.github/release-please.yml @@ -1,2 +1,14 @@ releaseType: python handleGHRelease: true +# NOTE: this section is generated by synthtool.languages.python +# See https://github.com/googleapis/synthtool/blob/master/synthtool/languages/python.py +branches: +- branch: v2 + handleGHRelease: true + releaseType: python +- branch: v1 + handleGHRelease: true + releaseType: python +- branch: v0 + handleGHRelease: true + releaseType: python diff --git a/.kokoro/docker/docs/Dockerfile b/.kokoro/docker/docs/Dockerfile index 4e1b1fb8..238b87b9 100644 --- a/.kokoro/docker/docs/Dockerfile +++ b/.kokoro/docker/docs/Dockerfile @@ -12,7 +12,7 @@ # See the License for the specific language governing permissions and # limitations under the License. -from ubuntu:20.04 +from ubuntu:22.04 ENV DEBIAN_FRONTEND noninteractive @@ -60,8 +60,24 @@ RUN apt-get update \ && rm -rf /var/lib/apt/lists/* \ && rm -f /var/cache/apt/archives/*.deb +###################### Install python 3.8.11 + +# Download python 3.8.11 +RUN wget https://www.python.org/ftp/python/3.8.11/Python-3.8.11.tgz + +# Extract files +RUN tar -xvf Python-3.8.11.tgz + +# Install python 3.8.11 +RUN ./Python-3.8.11/configure --enable-optimizations +RUN make altinstall + +###################### Install pip RUN wget -O /tmp/get-pip.py 'https://bootstrap.pypa.io/get-pip.py' \ - && python3.8 /tmp/get-pip.py \ + && python3 /tmp/get-pip.py \ && rm /tmp/get-pip.py +# Test pip +RUN python3 -m pip + CMD ["python3.8"] diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index 62eb5a77..46d23716 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -22,7 +22,7 @@ repos: - id: end-of-file-fixer - id: check-yaml - repo: https://github.com/psf/black - rev: 19.10b0 + rev: 22.3.0 hooks: - id: black - repo: https://gitlab.com/pycqa/flake8 diff --git a/CHANGELOG.md b/CHANGELOG.md index 16651ad6..07e1f84e 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,7 +4,19 @@ [1]: https://pypi.org/project/google-cloud-bigquery-datatransfer/#history -### [3.6.1](https://github.com/googleapis/python-bigquery-datatransfer/compare/v3.6.0...v3.6.1) (2022-03-05) +## [3.6.2](https://github.com/googleapis/python-bigquery-datatransfer/compare/v3.6.1...v3.6.2) (2022-06-03) + + +### Bug Fixes + +* **deps:** require protobuf <4.0.0dev ([#349](https://github.com/googleapis/python-bigquery-datatransfer/issues/349)) ([6b03c4b](https://github.com/googleapis/python-bigquery-datatransfer/commit/6b03c4b44b391f4cb9c47cebe40716db66a91ca1)) + + +### Documentation + +* fix changelog header to consistent size ([#350](https://github.com/googleapis/python-bigquery-datatransfer/issues/350)) ([c18147a](https://github.com/googleapis/python-bigquery-datatransfer/commit/c18147ab9d3713736ccdf7f327f1913c9be1e994)) + +## [3.6.1](https://github.com/googleapis/python-bigquery-datatransfer/compare/v3.6.0...v3.6.1) (2022-03-05) ### Bug Fixes @@ -40,7 +52,7 @@ * add support for Python 3.9 / 3.10 ([#253](https://github.com/googleapis/python-bigquery-datatransfer/issues/253)) ([5e02bd1](https://github.com/googleapis/python-bigquery-datatransfer/commit/5e02bd166149b4ef7ee0f295fdcb6a3570ec2f58)) -### [3.4.1](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.4.0...v3.4.1) (2021-11-01) +## [3.4.1](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.4.0...v3.4.1) (2021-11-01) ### Bug Fixes @@ -60,28 +72,28 @@ * add context manager support in client ([#234](https://www.github.com/googleapis/python-bigquery-datatransfer/issues/234)) ([cba3dad](https://www.github.com/googleapis/python-bigquery-datatransfer/commit/cba3dad55db8a00f95503b9c94d4ca21ed462a4f)) -### [3.3.4](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.3.3...v3.3.4) (2021-10-04) +## [3.3.4](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.3.3...v3.3.4) (2021-10-04) ### Bug Fixes * improper types in pagers generation ([702a848](https://www.github.com/googleapis/python-bigquery-datatransfer/commit/702a84864e5395cf7a7b8f2aeed93d2109414f7c)) -### [3.3.3](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.3.2...v3.3.3) (2021-09-24) +## [3.3.3](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.3.2...v3.3.3) (2021-09-24) ### Bug Fixes * add 'dict' annotation type to 'request' ([fe024db](https://www.github.com/googleapis/python-bigquery-datatransfer/commit/fe024dbcc3075fd779358d434f37ec3ee3a0adf9)) -### [3.3.2](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.3.1...v3.3.2) (2021-08-27) +## [3.3.2](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.3.1...v3.3.2) (2021-08-27) ### Documentation * Improvements to various message and field descriptions ([#202](https://www.github.com/googleapis/python-bigquery-datatransfer/issues/202)) ([9684723](https://www.github.com/googleapis/python-bigquery-datatransfer/commit/9684723aed8c743d665ccac2b338770dd09fc94a)) -### [3.3.1](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.3.0...v3.3.1) (2021-07-26) +## [3.3.1](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.3.0...v3.3.1) (2021-07-26) ### Bug Fixes @@ -128,7 +140,7 @@ * omit mention of Python 2.7 in 'CONTRIBUTING.rst' ([#1127](https://www.github.com/googleapis/python-bigquery-datatransfer/issues/1127)) ([#164](https://www.github.com/googleapis/python-bigquery-datatransfer/issues/164)) ([2741e4f](https://www.github.com/googleapis/python-bigquery-datatransfer/commit/2741e4fb1d9074494872fafcec96d870b14b671d)), closes [#1126](https://www.github.com/googleapis/python-bigquery-datatransfer/issues/1126) -### [3.1.1](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.1.0...v3.1.1) (2021-04-07) +## [3.1.1](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.1.0...v3.1.1) (2021-04-07) ### Bug Fixes @@ -143,7 +155,7 @@ * add `client_cert_source_for_mtls` parameter to grpc transport ([#114](https://www.github.com/googleapis/python-bigquery-datatransfer/issues/114)) ([9c7ed28](https://www.github.com/googleapis/python-bigquery-datatransfer/commit/9c7ed285203cd44917911e78c2c313d3ad50afda)) -### [3.0.1](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.0.0...v3.0.1) (2021-03-03) +## [3.0.1](https://www.github.com/googleapis/python-bigquery-datatransfer/compare/v3.0.0...v3.0.1) (2021-03-03) ### Bug Fixes diff --git a/docs/conf.py b/docs/conf.py index d55d2469..d037244d 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -24,9 +24,9 @@ # All configuration values have a default; values that are commented out # serve to show the default. -import sys import os import shlex +import sys # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the @@ -361,7 +361,10 @@ intersphinx_mapping = { "python": ("https://python.readthedocs.org/en/latest/", None), "google-auth": ("https://googleapis.dev/python/google-auth/latest/", None), - "google.api_core": ("https://googleapis.dev/python/google-api-core/latest/", None,), + "google.api_core": ( + "https://googleapis.dev/python/google-api-core/latest/", + None, + ), "grpc": ("https://grpc.github.io/grpc/python/", None), "proto-plus": ("https://proto-plus-python.readthedocs.io/en/latest/", None), "protobuf": ("https://googleapis.dev/python/protobuf/latest/", None), diff --git a/google/cloud/bigquery_datatransfer/__init__.py b/google/cloud/bigquery_datatransfer/__init__.py index ea8dc599..a02fe531 100644 --- a/google/cloud/bigquery_datatransfer/__init__.py +++ b/google/cloud/bigquery_datatransfer/__init__.py @@ -14,89 +14,48 @@ # limitations under the License. # -from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service.client import ( - DataTransferServiceClient, -) from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service.async_client import ( DataTransferServiceAsyncClient, ) - -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( - CheckValidCredsRequest, +from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service.client import ( + DataTransferServiceClient, ) from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( + CheckValidCredsRequest, CheckValidCredsResponse, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( CreateTransferConfigRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import DataSource -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import DataSourceParameter -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( + DataSource, + DataSourceParameter, DeleteTransferConfigRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( DeleteTransferRunRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( EnrollDataSourcesRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( GetDataSourceRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( GetTransferConfigRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( GetTransferRunRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( ListDataSourcesRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( ListDataSourcesResponse, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( ListTransferConfigsRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( ListTransferConfigsResponse, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( ListTransferLogsRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( ListTransferLogsResponse, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( ListTransferRunsRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( ListTransferRunsResponse, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( ScheduleTransferRunsRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( ScheduleTransferRunsResponse, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( StartManualTransferRunsRequest, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( StartManualTransferRunsResponse, -) -from google.cloud.bigquery_datatransfer_v1.types.datatransfer import ( UpdateTransferConfigRequest, ) -from google.cloud.bigquery_datatransfer_v1.types.transfer import EmailPreferences -from google.cloud.bigquery_datatransfer_v1.types.transfer import ScheduleOptions -from google.cloud.bigquery_datatransfer_v1.types.transfer import TransferConfig -from google.cloud.bigquery_datatransfer_v1.types.transfer import TransferMessage -from google.cloud.bigquery_datatransfer_v1.types.transfer import TransferRun -from google.cloud.bigquery_datatransfer_v1.types.transfer import UserInfo -from google.cloud.bigquery_datatransfer_v1.types.transfer import TransferState -from google.cloud.bigquery_datatransfer_v1.types.transfer import TransferType +from google.cloud.bigquery_datatransfer_v1.types.transfer import ( + EmailPreferences, + ScheduleOptions, + TransferConfig, + TransferMessage, + TransferRun, + TransferState, + TransferType, + UserInfo, +) __all__ = ( "DataTransferServiceClient", diff --git a/google/cloud/bigquery_datatransfer_v1/__init__.py b/google/cloud/bigquery_datatransfer_v1/__init__.py index cbc23487..30d7482e 100644 --- a/google/cloud/bigquery_datatransfer_v1/__init__.py +++ b/google/cloud/bigquery_datatransfer_v1/__init__.py @@ -14,41 +14,46 @@ # limitations under the License. # -from .services.data_transfer_service import DataTransferServiceClient -from .services.data_transfer_service import DataTransferServiceAsyncClient - -from .types.datatransfer import CheckValidCredsRequest -from .types.datatransfer import CheckValidCredsResponse -from .types.datatransfer import CreateTransferConfigRequest -from .types.datatransfer import DataSource -from .types.datatransfer import DataSourceParameter -from .types.datatransfer import DeleteTransferConfigRequest -from .types.datatransfer import DeleteTransferRunRequest -from .types.datatransfer import EnrollDataSourcesRequest -from .types.datatransfer import GetDataSourceRequest -from .types.datatransfer import GetTransferConfigRequest -from .types.datatransfer import GetTransferRunRequest -from .types.datatransfer import ListDataSourcesRequest -from .types.datatransfer import ListDataSourcesResponse -from .types.datatransfer import ListTransferConfigsRequest -from .types.datatransfer import ListTransferConfigsResponse -from .types.datatransfer import ListTransferLogsRequest -from .types.datatransfer import ListTransferLogsResponse -from .types.datatransfer import ListTransferRunsRequest -from .types.datatransfer import ListTransferRunsResponse -from .types.datatransfer import ScheduleTransferRunsRequest -from .types.datatransfer import ScheduleTransferRunsResponse -from .types.datatransfer import StartManualTransferRunsRequest -from .types.datatransfer import StartManualTransferRunsResponse -from .types.datatransfer import UpdateTransferConfigRequest -from .types.transfer import EmailPreferences -from .types.transfer import ScheduleOptions -from .types.transfer import TransferConfig -from .types.transfer import TransferMessage -from .types.transfer import TransferRun -from .types.transfer import UserInfo -from .types.transfer import TransferState -from .types.transfer import TransferType +from .services.data_transfer_service import ( + DataTransferServiceAsyncClient, + DataTransferServiceClient, +) +from .types.datatransfer import ( + CheckValidCredsRequest, + CheckValidCredsResponse, + CreateTransferConfigRequest, + DataSource, + DataSourceParameter, + DeleteTransferConfigRequest, + DeleteTransferRunRequest, + EnrollDataSourcesRequest, + GetDataSourceRequest, + GetTransferConfigRequest, + GetTransferRunRequest, + ListDataSourcesRequest, + ListDataSourcesResponse, + ListTransferConfigsRequest, + ListTransferConfigsResponse, + ListTransferLogsRequest, + ListTransferLogsResponse, + ListTransferRunsRequest, + ListTransferRunsResponse, + ScheduleTransferRunsRequest, + ScheduleTransferRunsResponse, + StartManualTransferRunsRequest, + StartManualTransferRunsResponse, + UpdateTransferConfigRequest, +) +from .types.transfer import ( + EmailPreferences, + ScheduleOptions, + TransferConfig, + TransferMessage, + TransferRun, + TransferState, + TransferType, + UserInfo, +) __all__ = ( "DataTransferServiceAsyncClient", diff --git a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/__init__.py b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/__init__.py index 4cdbdc6b..822a5513 100644 --- a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/__init__.py +++ b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/__init__.py @@ -13,8 +13,8 @@ # See the License for the specific language governing permissions and # limitations under the License. # -from .client import DataTransferServiceClient from .async_client import DataTransferServiceAsyncClient +from .client import DataTransferServiceClient __all__ = ( "DataTransferServiceClient", diff --git a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/async_client.py b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/async_client.py index 9f0c5397..29b53e20 100644 --- a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/async_client.py +++ b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/async_client.py @@ -16,33 +16,34 @@ from collections import OrderedDict import functools import re -from typing import Dict, Optional, Sequence, Tuple, Type, Union -import pkg_resources +from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union import warnings -from google.api_core.client_options import ClientOptions from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 from google.api_core import retry as retries +from google.api_core.client_options import ClientOptions from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore +import pkg_resources try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] except AttributeError: # pragma: NO COVER OptionalRetry = Union[retries.Retry, object] # type: ignore -from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service import pagers -from google.cloud.bigquery_datatransfer_v1.types import datatransfer -from google.cloud.bigquery_datatransfer_v1.types import transfer from google.protobuf import duration_pb2 # type: ignore from google.protobuf import field_mask_pb2 # type: ignore from google.protobuf import struct_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore from google.rpc import status_pb2 # type: ignore -from .transports.base import DataTransferServiceTransport, DEFAULT_CLIENT_INFO -from .transports.grpc_asyncio import DataTransferServiceGrpcAsyncIOTransport + +from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service import pagers +from google.cloud.bigquery_datatransfer_v1.types import datatransfer, transfer + from .client import DataTransferServiceClient +from .transports.base import DEFAULT_CLIENT_INFO, DataTransferServiceTransport +from .transports.grpc_asyncio import DataTransferServiceGrpcAsyncIOTransport class DataTransferServiceAsyncClient: @@ -232,14 +233,13 @@ async def get_data_source( r"""Retrieves a supported data source and returns its settings. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 - def sample_get_data_source(): + async def sample_get_data_source(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.GetDataSourceRequest( @@ -247,7 +247,7 @@ def sample_get_data_source(): ) # Make the request - response = client.get_data_source(request=request) + response = await client.get_data_source(request=request) # Handle the response print(response) @@ -319,7 +319,12 @@ def sample_get_data_source(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -336,14 +341,13 @@ async def list_data_sources( r"""Lists supported data sources and returns their settings. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 - def sample_list_data_sources(): + async def sample_list_data_sources(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.ListDataSourcesRequest( @@ -354,7 +358,7 @@ def sample_list_data_sources(): page_result = client.list_data_sources(request=request) # Handle the response - for response in page_result: + async for response in page_result: print(response) Args: @@ -427,12 +431,20 @@ def sample_list_data_sources(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__aiter__` convenience method. response = pagers.ListDataSourcesAsyncPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -454,9 +466,9 @@ async def create_transfer_config( from google.cloud import bigquery_datatransfer_v1 - def sample_create_transfer_config(): + async def sample_create_transfer_config(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) transfer_config = bigquery_datatransfer_v1.TransferConfig() @@ -468,7 +480,7 @@ def sample_create_transfer_config(): ) # Make the request - response = client.create_transfer_config(request=request) + response = await client.create_transfer_config(request=request) # Handle the response print(response) @@ -553,7 +565,12 @@ def sample_create_transfer_config(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -571,14 +588,13 @@ async def update_transfer_config( r"""Updates a data transfer configuration. All fields must be set, even if they are not updated. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 - def sample_update_transfer_config(): + async def sample_update_transfer_config(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) transfer_config = bigquery_datatransfer_v1.TransferConfig() @@ -589,7 +605,7 @@ def sample_update_transfer_config(): ) # Make the request - response = client.update_transfer_config(request=request) + response = await client.update_transfer_config(request=request) # Handle the response print(response) @@ -668,7 +684,12 @@ def sample_update_transfer_config(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -685,14 +706,13 @@ async def delete_transfer_config( r"""Deletes a data transfer configuration, including any associated transfer runs and logs. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 - def sample_delete_transfer_config(): + async def sample_delete_transfer_config(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.DeleteTransferConfigRequest( @@ -700,7 +720,7 @@ def sample_delete_transfer_config(): ) # Make the request - client.delete_transfer_config(request=request) + await client.delete_transfer_config(request=request) Args: request (Union[google.cloud.bigquery_datatransfer_v1.types.DeleteTransferConfigRequest, dict]): @@ -765,7 +785,10 @@ def sample_delete_transfer_config(): # Send the request. await rpc( - request, retry=retry, timeout=timeout, metadata=metadata, + request, + retry=retry, + timeout=timeout, + metadata=metadata, ) async def get_transfer_config( @@ -783,9 +806,9 @@ async def get_transfer_config( from google.cloud import bigquery_datatransfer_v1 - def sample_get_transfer_config(): + async def sample_get_transfer_config(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.GetTransferConfigRequest( @@ -793,7 +816,7 @@ def sample_get_transfer_config(): ) # Make the request - response = client.get_transfer_config(request=request) + response = await client.get_transfer_config(request=request) # Handle the response print(response) @@ -871,7 +894,12 @@ def sample_get_transfer_config(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -888,14 +916,13 @@ async def list_transfer_configs( r"""Returns information about all transfer configs owned by a project in the specified location. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 - def sample_list_transfer_configs(): + async def sample_list_transfer_configs(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.ListTransferConfigsRequest( @@ -906,7 +933,7 @@ def sample_list_transfer_configs(): page_result = client.list_transfer_configs(request=request) # Handle the response - for response in page_result: + async for response in page_result: print(response) Args: @@ -978,12 +1005,20 @@ def sample_list_transfer_configs(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__aiter__` convenience method. response = pagers.ListTransferConfigsAsyncPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -1006,14 +1041,13 @@ async def schedule_transfer_runs( created per UTC time in the time range. DEPRECATED: use StartManualTransferRuns instead. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 - def sample_schedule_transfer_runs(): + async def sample_schedule_transfer_runs(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.ScheduleTransferRunsRequest( @@ -1021,7 +1055,7 @@ def sample_schedule_transfer_runs(): ) # Make the request - response = client.schedule_transfer_runs(request=request) + response = await client.schedule_transfer_runs(request=request) # Handle the response print(response) @@ -1105,7 +1139,12 @@ def sample_schedule_transfer_runs(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1123,21 +1162,20 @@ async def start_manual_transfer_runs( time range where the run_time is between start_time (inclusive) and end_time (exclusive), or for a specific run_time. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 - def sample_start_manual_transfer_runs(): + async def sample_start_manual_transfer_runs(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.StartManualTransferRunsRequest( ) # Make the request - response = client.start_manual_transfer_runs(request=request) + response = await client.start_manual_transfer_runs(request=request) # Handle the response print(response) @@ -1176,7 +1214,12 @@ def sample_start_manual_transfer_runs(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1193,14 +1236,13 @@ async def get_transfer_run( r"""Returns information about the particular transfer run. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 - def sample_get_transfer_run(): + async def sample_get_transfer_run(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.GetTransferRunRequest( @@ -1208,7 +1250,7 @@ def sample_get_transfer_run(): ) # Make the request - response = client.get_transfer_run(request=request) + response = await client.get_transfer_run(request=request) # Handle the response print(response) @@ -1279,7 +1321,12 @@ def sample_get_transfer_run(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1299,9 +1346,9 @@ async def delete_transfer_run( from google.cloud import bigquery_datatransfer_v1 - def sample_delete_transfer_run(): + async def sample_delete_transfer_run(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.DeleteTransferRunRequest( @@ -1309,7 +1356,7 @@ def sample_delete_transfer_run(): ) # Make the request - client.delete_transfer_run(request=request) + await client.delete_transfer_run(request=request) Args: request (Union[google.cloud.bigquery_datatransfer_v1.types.DeleteTransferRunRequest, dict]): @@ -1374,7 +1421,10 @@ def sample_delete_transfer_run(): # Send the request. await rpc( - request, retry=retry, timeout=timeout, metadata=metadata, + request, + retry=retry, + timeout=timeout, + metadata=metadata, ) async def list_transfer_runs( @@ -1389,14 +1439,13 @@ async def list_transfer_runs( r"""Returns information about running and completed transfer runs. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 - def sample_list_transfer_runs(): + async def sample_list_transfer_runs(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.ListTransferRunsRequest( @@ -1407,7 +1456,7 @@ def sample_list_transfer_runs(): page_result = client.list_transfer_runs(request=request) # Handle the response - for response in page_result: + async for response in page_result: print(response) Args: @@ -1481,12 +1530,20 @@ def sample_list_transfer_runs(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__aiter__` convenience method. response = pagers.ListTransferRunsAsyncPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -1507,9 +1564,9 @@ async def list_transfer_logs( from google.cloud import bigquery_datatransfer_v1 - def sample_list_transfer_logs(): + async def sample_list_transfer_logs(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.ListTransferLogsRequest( @@ -1520,7 +1577,7 @@ def sample_list_transfer_logs(): page_result = client.list_transfer_logs(request=request) # Handle the response - for response in page_result: + async for response in page_result: print(response) Args: @@ -1593,12 +1650,20 @@ def sample_list_transfer_logs(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__aiter__` convenience method. response = pagers.ListTransferLogsAsyncPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -1616,14 +1681,13 @@ async def check_valid_creds( r"""Returns true if valid credentials exist for the given data source and requesting user. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 - def sample_check_valid_creds(): + async def sample_check_valid_creds(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.CheckValidCredsRequest( @@ -1631,7 +1695,7 @@ def sample_check_valid_creds(): ) # Make the request - response = client.check_valid_creds(request=request) + response = await client.check_valid_creds(request=request) # Handle the response print(response) @@ -1709,7 +1773,12 @@ def sample_check_valid_creds(): ) # Send the request. - response = await rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = await rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1731,21 +1800,20 @@ async def enroll_data_sources( https://cloud.google.com/bigquery/bigquery-web-ui and https://cloud.google.com/bigquery/docs/working-with-transfers). - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 - def sample_enroll_data_sources(): + async def sample_enroll_data_sources(): # Create a client - client = bigquery_datatransfer_v1.DataTransferServiceClient() + client = bigquery_datatransfer_v1.DataTransferServiceAsyncClient() # Initialize request argument(s) request = bigquery_datatransfer_v1.EnrollDataSourcesRequest( ) # Make the request - client.enroll_data_sources(request=request) + await client.enroll_data_sources(request=request) Args: request (Union[google.cloud.bigquery_datatransfer_v1.types.EnrollDataSourcesRequest, dict]): @@ -1777,7 +1845,10 @@ def sample_enroll_data_sources(): # Send the request. await rpc( - request, retry=retry, timeout=timeout, metadata=metadata, + request, + retry=retry, + timeout=timeout, + metadata=metadata, ) async def __aenter__(self): diff --git a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/client.py b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/client.py index 44438bb4..362ee4a5 100644 --- a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/client.py +++ b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/client.py @@ -16,8 +16,7 @@ from collections import OrderedDict import os import re -from typing import Dict, Optional, Sequence, Tuple, Type, Union -import pkg_resources +from typing import Dict, Mapping, Optional, Sequence, Tuple, Type, Union import warnings from google.api_core import client_options as client_options_lib @@ -25,25 +24,27 @@ from google.api_core import gapic_v1 from google.api_core import retry as retries from google.auth import credentials as ga_credentials # type: ignore +from google.auth.exceptions import MutualTLSChannelError # type: ignore from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore -from google.auth.exceptions import MutualTLSChannelError # type: ignore from google.oauth2 import service_account # type: ignore +import pkg_resources try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] except AttributeError: # pragma: NO COVER OptionalRetry = Union[retries.Retry, object] # type: ignore -from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service import pagers -from google.cloud.bigquery_datatransfer_v1.types import datatransfer -from google.cloud.bigquery_datatransfer_v1.types import transfer from google.protobuf import duration_pb2 # type: ignore from google.protobuf import field_mask_pb2 # type: ignore from google.protobuf import struct_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore from google.rpc import status_pb2 # type: ignore -from .transports.base import DataTransferServiceTransport, DEFAULT_CLIENT_INFO + +from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service import pagers +from google.cloud.bigquery_datatransfer_v1.types import datatransfer, transfer + +from .transports.base import DEFAULT_CLIENT_INFO, DataTransferServiceTransport from .transports.grpc import DataTransferServiceGrpcTransport from .transports.grpc_asyncio import DataTransferServiceGrpcAsyncIOTransport @@ -63,7 +64,8 @@ class DataTransferServiceClientMeta(type): _transport_registry["grpc_asyncio"] = DataTransferServiceGrpcAsyncIOTransport def get_transport_class( - cls, label: str = None, + cls, + label: str = None, ) -> Type[DataTransferServiceTransport]: """Returns an appropriate transport class. @@ -171,10 +173,14 @@ def transport(self) -> DataTransferServiceTransport: return self._transport @staticmethod - def data_source_path(project: str, data_source: str,) -> str: + def data_source_path( + project: str, + data_source: str, + ) -> str: """Returns a fully-qualified data_source string.""" return "projects/{project}/dataSources/{data_source}".format( - project=project, data_source=data_source, + project=project, + data_source=data_source, ) @staticmethod @@ -186,10 +192,16 @@ def parse_data_source_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def run_path(project: str, transfer_config: str, run: str,) -> str: + def run_path( + project: str, + transfer_config: str, + run: str, + ) -> str: """Returns a fully-qualified run string.""" return "projects/{project}/transferConfigs/{transfer_config}/runs/{run}".format( - project=project, transfer_config=transfer_config, run=run, + project=project, + transfer_config=transfer_config, + run=run, ) @staticmethod @@ -202,10 +214,14 @@ def parse_run_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def transfer_config_path(project: str, transfer_config: str,) -> str: + def transfer_config_path( + project: str, + transfer_config: str, + ) -> str: """Returns a fully-qualified transfer_config string.""" return "projects/{project}/transferConfigs/{transfer_config}".format( - project=project, transfer_config=transfer_config, + project=project, + transfer_config=transfer_config, ) @staticmethod @@ -218,7 +234,9 @@ def parse_transfer_config_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_billing_account_path(billing_account: str,) -> str: + def common_billing_account_path( + billing_account: str, + ) -> str: """Returns a fully-qualified billing_account string.""" return "billingAccounts/{billing_account}".format( billing_account=billing_account, @@ -231,9 +249,13 @@ def parse_common_billing_account_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_folder_path(folder: str,) -> str: + def common_folder_path( + folder: str, + ) -> str: """Returns a fully-qualified folder string.""" - return "folders/{folder}".format(folder=folder,) + return "folders/{folder}".format( + folder=folder, + ) @staticmethod def parse_common_folder_path(path: str) -> Dict[str, str]: @@ -242,9 +264,13 @@ def parse_common_folder_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_organization_path(organization: str,) -> str: + def common_organization_path( + organization: str, + ) -> str: """Returns a fully-qualified organization string.""" - return "organizations/{organization}".format(organization=organization,) + return "organizations/{organization}".format( + organization=organization, + ) @staticmethod def parse_common_organization_path(path: str) -> Dict[str, str]: @@ -253,9 +279,13 @@ def parse_common_organization_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_project_path(project: str,) -> str: + def common_project_path( + project: str, + ) -> str: """Returns a fully-qualified project string.""" - return "projects/{project}".format(project=project,) + return "projects/{project}".format( + project=project, + ) @staticmethod def parse_common_project_path(path: str) -> Dict[str, str]: @@ -264,10 +294,14 @@ def parse_common_project_path(path: str) -> Dict[str, str]: return m.groupdict() if m else {} @staticmethod - def common_location_path(project: str, location: str,) -> str: + def common_location_path( + project: str, + location: str, + ) -> str: """Returns a fully-qualified location string.""" return "projects/{project}/locations/{location}".format( - project=project, location=location, + project=project, + location=location, ) @staticmethod @@ -453,7 +487,6 @@ def get_data_source( r"""Retrieves a supported data source and returns its settings. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 @@ -530,7 +563,12 @@ def sample_get_data_source(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -547,7 +585,6 @@ def list_data_sources( r"""Lists supported data sources and returns their settings. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 @@ -628,12 +665,20 @@ def sample_list_data_sources(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__iter__` convenience method. response = pagers.ListDataSourcesPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -754,7 +799,12 @@ def sample_create_transfer_config(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -772,7 +822,6 @@ def update_transfer_config( r"""Updates a data transfer configuration. All fields must be set, even if they are not updated. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 @@ -869,7 +918,12 @@ def sample_update_transfer_config(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -886,7 +940,6 @@ def delete_transfer_config( r"""Deletes a data transfer configuration, including any associated transfer runs and logs. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 @@ -956,7 +1009,10 @@ def sample_delete_transfer_config(): # Send the request. rpc( - request, retry=retry, timeout=timeout, metadata=metadata, + request, + retry=retry, + timeout=timeout, + metadata=metadata, ) def get_transfer_config( @@ -1052,7 +1108,12 @@ def sample_get_transfer_config(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1069,7 +1130,6 @@ def list_transfer_configs( r"""Returns information about all transfer configs owned by a project in the specified location. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 @@ -1149,12 +1209,20 @@ def sample_list_transfer_configs(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__iter__` convenience method. response = pagers.ListTransferConfigsPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -1177,7 +1245,6 @@ def schedule_transfer_runs( created per UTC time in the time range. DEPRECATED: use StartManualTransferRuns instead. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 @@ -1276,7 +1343,12 @@ def sample_schedule_transfer_runs(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1294,7 +1366,6 @@ def start_manual_transfer_runs( time range where the run_time is between start_time (inclusive) and end_time (exclusive), or for a specific run_time. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 @@ -1350,7 +1421,12 @@ def sample_start_manual_transfer_runs(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1367,7 +1443,6 @@ def get_transfer_run( r"""Returns information about the particular transfer run. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 @@ -1443,7 +1518,12 @@ def sample_get_transfer_run(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1528,7 +1608,10 @@ def sample_delete_transfer_run(): # Send the request. rpc( - request, retry=retry, timeout=timeout, metadata=metadata, + request, + retry=retry, + timeout=timeout, + metadata=metadata, ) def list_transfer_runs( @@ -1543,7 +1626,6 @@ def list_transfer_runs( r"""Returns information about running and completed transfer runs. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 @@ -1625,12 +1707,20 @@ def sample_list_transfer_runs(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__iter__` convenience method. response = pagers.ListTransferRunsPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -1727,12 +1817,20 @@ def sample_list_transfer_logs(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # This method is paged; wrap the response in a pager, which provides # an `__iter__` convenience method. response = pagers.ListTransferLogsPager( - method=rpc, request=request, response=response, metadata=metadata, + method=rpc, + request=request, + response=response, + metadata=metadata, ) # Done; return the response. @@ -1750,7 +1848,6 @@ def check_valid_creds( r"""Returns true if valid credentials exist for the given data source and requesting user. - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 @@ -1833,7 +1930,12 @@ def sample_check_valid_creds(): ) # Send the request. - response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) + response = rpc( + request, + retry=retry, + timeout=timeout, + metadata=metadata, + ) # Done; return the response. return response @@ -1855,7 +1957,6 @@ def enroll_data_sources( https://cloud.google.com/bigquery/bigquery-web-ui and https://cloud.google.com/bigquery/docs/working-with-transfers). - .. code-block:: python from google.cloud import bigquery_datatransfer_v1 @@ -1902,7 +2003,10 @@ def sample_enroll_data_sources(): # Send the request. rpc( - request, retry=retry, timeout=timeout, metadata=metadata, + request, + retry=retry, + timeout=timeout, + metadata=metadata, ) def __enter__(self): diff --git a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/pagers.py b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/pagers.py index 3cf2af5c..b22f400a 100644 --- a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/pagers.py +++ b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/pagers.py @@ -18,14 +18,13 @@ AsyncIterator, Awaitable, Callable, + Iterator, + Optional, Sequence, Tuple, - Optional, - Iterator, ) -from google.cloud.bigquery_datatransfer_v1.types import datatransfer -from google.cloud.bigquery_datatransfer_v1.types import transfer +from google.cloud.bigquery_datatransfer_v1.types import datatransfer, transfer class ListDataSourcesPager: diff --git a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/__init__.py b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/__init__.py index c2e85d32..01d735cb 100644 --- a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/__init__.py +++ b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/__init__.py @@ -20,7 +20,6 @@ from .grpc import DataTransferServiceGrpcTransport from .grpc_asyncio import DataTransferServiceGrpcAsyncIOTransport - # Compile a registry of transports. _transport_registry = ( OrderedDict() diff --git a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/base.py b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/base.py index 83261a2b..408ec8c8 100644 --- a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/base.py +++ b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/base.py @@ -15,19 +15,18 @@ # import abc from typing import Awaitable, Callable, Dict, Optional, Sequence, Union -import pkg_resources -import google.auth # type: ignore import google.api_core from google.api_core import exceptions as core_exceptions from google.api_core import gapic_v1 from google.api_core import retry as retries +import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.oauth2 import service_account # type: ignore - -from google.cloud.bigquery_datatransfer_v1.types import datatransfer -from google.cloud.bigquery_datatransfer_v1.types import transfer from google.protobuf import empty_pb2 # type: ignore +import pkg_resources + +from google.cloud.bigquery_datatransfer_v1.types import datatransfer, transfer try: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( @@ -82,6 +81,7 @@ def __init__( always_use_jwt_access (Optional[bool]): Whether self signed JWT should be used for service account credentials. """ + # Save the hostname. Default to port 443 (HTTPS) if none is specified. if ":" not in host: host += ":443" @@ -293,16 +293,18 @@ def _prep_wrapped_messages(self, client_info): client_info=client_info, ), self.enroll_data_sources: gapic_v1.method.wrap_method( - self.enroll_data_sources, default_timeout=None, client_info=client_info, + self.enroll_data_sources, + default_timeout=None, + client_info=client_info, ), } def close(self): """Closes resources associated with the transport. - .. warning:: - Only call this method if the transport is NOT shared - with other clients - this may cause errors in other clients! + .. warning:: + Only call this method if the transport is NOT shared + with other clients - this may cause errors in other clients! """ raise NotImplementedError() @@ -462,5 +464,9 @@ def enroll_data_sources( ]: raise NotImplementedError() + @property + def kind(self) -> str: + raise NotImplementedError() + __all__ = ("DataTransferServiceTransport",) diff --git a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/grpc.py b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/grpc.py index e6ed06be..5cf8b3b5 100644 --- a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/grpc.py +++ b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/grpc.py @@ -13,21 +13,19 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import warnings from typing import Callable, Dict, Optional, Sequence, Tuple, Union +import warnings -from google.api_core import grpc_helpers -from google.api_core import gapic_v1 +from google.api_core import gapic_v1, grpc_helpers import google.auth # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore - +from google.protobuf import empty_pb2 # type: ignore import grpc # type: ignore -from google.cloud.bigquery_datatransfer_v1.types import datatransfer -from google.cloud.bigquery_datatransfer_v1.types import transfer -from google.protobuf import empty_pb2 # type: ignore -from .base import DataTransferServiceTransport, DEFAULT_CLIENT_INFO +from google.cloud.bigquery_datatransfer_v1.types import datatransfer, transfer + +from .base import DEFAULT_CLIENT_INFO, DataTransferServiceTransport class DataTransferServiceGrpcTransport(DataTransferServiceTransport): @@ -227,8 +225,7 @@ def create_channel( @property def grpc_channel(self) -> grpc.Channel: - """Return the channel designed to connect to this service. - """ + """Return the channel designed to connect to this service.""" return self._grpc_channel @property @@ -663,5 +660,9 @@ def enroll_data_sources( def close(self): self.grpc_channel.close() + @property + def kind(self) -> str: + return "grpc" + __all__ = ("DataTransferServiceGrpcTransport",) diff --git a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/grpc_asyncio.py b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/grpc_asyncio.py index ab4ecb3e..4e204ea5 100644 --- a/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/grpc_asyncio.py +++ b/google/cloud/bigquery_datatransfer_v1/services/data_transfer_service/transports/grpc_asyncio.py @@ -13,21 +13,19 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import warnings from typing import Awaitable, Callable, Dict, Optional, Sequence, Tuple, Union +import warnings -from google.api_core import gapic_v1 -from google.api_core import grpc_helpers_async +from google.api_core import gapic_v1, grpc_helpers_async from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore - +from google.protobuf import empty_pb2 # type: ignore import grpc # type: ignore from grpc.experimental import aio # type: ignore -from google.cloud.bigquery_datatransfer_v1.types import datatransfer -from google.cloud.bigquery_datatransfer_v1.types import transfer -from google.protobuf import empty_pb2 # type: ignore -from .base import DataTransferServiceTransport, DEFAULT_CLIENT_INFO +from google.cloud.bigquery_datatransfer_v1.types import datatransfer, transfer + +from .base import DEFAULT_CLIENT_INFO, DataTransferServiceTransport from .grpc import DataTransferServiceGrpcTransport diff --git a/google/cloud/bigquery_datatransfer_v1/types/__init__.py b/google/cloud/bigquery_datatransfer_v1/types/__init__.py index 43094514..0e47e7cf 100644 --- a/google/cloud/bigquery_datatransfer_v1/types/__init__.py +++ b/google/cloud/bigquery_datatransfer_v1/types/__init__.py @@ -45,9 +45,9 @@ TransferConfig, TransferMessage, TransferRun, - UserInfo, TransferState, TransferType, + UserInfo, ) __all__ = ( diff --git a/google/cloud/bigquery_datatransfer_v1/types/datatransfer.py b/google/cloud/bigquery_datatransfer_v1/types/datatransfer.py index cd731557..443678f2 100644 --- a/google/cloud/bigquery_datatransfer_v1/types/datatransfer.py +++ b/google/cloud/bigquery_datatransfer_v1/types/datatransfer.py @@ -13,14 +13,13 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import proto # type: ignore - -from google.cloud.bigquery_datatransfer_v1.types import transfer from google.protobuf import duration_pb2 # type: ignore from google.protobuf import field_mask_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore from google.protobuf import wrappers_pb2 # type: ignore +import proto # type: ignore +from google.cloud.bigquery_datatransfer_v1.types import transfer __protobuf__ = proto.module( package="google.cloud.bigquery.datatransfer.v1", @@ -110,24 +109,74 @@ class Type(proto.Enum): RECORD = 5 PLUS_PAGE = 6 - param_id = proto.Field(proto.STRING, number=1,) - display_name = proto.Field(proto.STRING, number=2,) - description = proto.Field(proto.STRING, number=3,) - type_ = proto.Field(proto.ENUM, number=4, enum=Type,) - required = proto.Field(proto.BOOL, number=5,) - repeated = proto.Field(proto.BOOL, number=6,) - validation_regex = proto.Field(proto.STRING, number=7,) - allowed_values = proto.RepeatedField(proto.STRING, number=8,) - min_value = proto.Field(proto.MESSAGE, number=9, message=wrappers_pb2.DoubleValue,) - max_value = proto.Field(proto.MESSAGE, number=10, message=wrappers_pb2.DoubleValue,) + param_id = proto.Field( + proto.STRING, + number=1, + ) + display_name = proto.Field( + proto.STRING, + number=2, + ) + description = proto.Field( + proto.STRING, + number=3, + ) + type_ = proto.Field( + proto.ENUM, + number=4, + enum=Type, + ) + required = proto.Field( + proto.BOOL, + number=5, + ) + repeated = proto.Field( + proto.BOOL, + number=6, + ) + validation_regex = proto.Field( + proto.STRING, + number=7, + ) + allowed_values = proto.RepeatedField( + proto.STRING, + number=8, + ) + min_value = proto.Field( + proto.MESSAGE, + number=9, + message=wrappers_pb2.DoubleValue, + ) + max_value = proto.Field( + proto.MESSAGE, + number=10, + message=wrappers_pb2.DoubleValue, + ) fields = proto.RepeatedField( - proto.MESSAGE, number=11, message="DataSourceParameter", + proto.MESSAGE, + number=11, + message="DataSourceParameter", + ) + validation_description = proto.Field( + proto.STRING, + number=12, + ) + validation_help_url = proto.Field( + proto.STRING, + number=13, + ) + immutable = proto.Field( + proto.BOOL, + number=14, + ) + recurse = proto.Field( + proto.BOOL, + number=15, + ) + deprecated = proto.Field( + proto.BOOL, + number=20, ) - validation_description = proto.Field(proto.STRING, number=12,) - validation_help_url = proto.Field(proto.STRING, number=13,) - immutable = proto.Field(proto.BOOL, number=14,) - recurse = proto.Field(proto.BOOL, number=15,) - deprecated = proto.Field(proto.BOOL, number=20,) class DataSource(proto.Message): @@ -207,27 +256,82 @@ class DataRefreshType(proto.Enum): SLIDING_WINDOW = 1 CUSTOM_SLIDING_WINDOW = 2 - name = proto.Field(proto.STRING, number=1,) - data_source_id = proto.Field(proto.STRING, number=2,) - display_name = proto.Field(proto.STRING, number=3,) - description = proto.Field(proto.STRING, number=4,) - client_id = proto.Field(proto.STRING, number=5,) - scopes = proto.RepeatedField(proto.STRING, number=6,) - transfer_type = proto.Field(proto.ENUM, number=7, enum=transfer.TransferType,) - supports_multiple_transfers = proto.Field(proto.BOOL, number=8,) - update_deadline_seconds = proto.Field(proto.INT32, number=9,) - default_schedule = proto.Field(proto.STRING, number=10,) - supports_custom_schedule = proto.Field(proto.BOOL, number=11,) + name = proto.Field( + proto.STRING, + number=1, + ) + data_source_id = proto.Field( + proto.STRING, + number=2, + ) + display_name = proto.Field( + proto.STRING, + number=3, + ) + description = proto.Field( + proto.STRING, + number=4, + ) + client_id = proto.Field( + proto.STRING, + number=5, + ) + scopes = proto.RepeatedField( + proto.STRING, + number=6, + ) + transfer_type = proto.Field( + proto.ENUM, + number=7, + enum=transfer.TransferType, + ) + supports_multiple_transfers = proto.Field( + proto.BOOL, + number=8, + ) + update_deadline_seconds = proto.Field( + proto.INT32, + number=9, + ) + default_schedule = proto.Field( + proto.STRING, + number=10, + ) + supports_custom_schedule = proto.Field( + proto.BOOL, + number=11, + ) parameters = proto.RepeatedField( - proto.MESSAGE, number=12, message="DataSourceParameter", + proto.MESSAGE, + number=12, + message="DataSourceParameter", + ) + help_url = proto.Field( + proto.STRING, + number=13, + ) + authorization_type = proto.Field( + proto.ENUM, + number=14, + enum=AuthorizationType, + ) + data_refresh_type = proto.Field( + proto.ENUM, + number=15, + enum=DataRefreshType, + ) + default_data_refresh_window_days = proto.Field( + proto.INT32, + number=16, + ) + manual_runs_disabled = proto.Field( + proto.BOOL, + number=17, ) - help_url = proto.Field(proto.STRING, number=13,) - authorization_type = proto.Field(proto.ENUM, number=14, enum=AuthorizationType,) - data_refresh_type = proto.Field(proto.ENUM, number=15, enum=DataRefreshType,) - default_data_refresh_window_days = proto.Field(proto.INT32, number=16,) - manual_runs_disabled = proto.Field(proto.BOOL, number=17,) minimum_schedule_interval = proto.Field( - proto.MESSAGE, number=18, message=duration_pb2.Duration, + proto.MESSAGE, + number=18, + message=duration_pb2.Duration, ) @@ -242,7 +346,10 @@ class GetDataSourceRequest(proto.Message): ``projects/{project_id}/locations/{location_id}/dataSources/{data_source_id}`` """ - name = proto.Field(proto.STRING, number=1,) + name = proto.Field( + proto.STRING, + number=1, + ) class ListDataSourcesRequest(proto.Message): @@ -266,9 +373,18 @@ class ListDataSourcesRequest(proto.Message): maximum value of 1000 results. """ - parent = proto.Field(proto.STRING, number=1,) - page_token = proto.Field(proto.STRING, number=3,) - page_size = proto.Field(proto.INT32, number=4,) + parent = proto.Field( + proto.STRING, + number=1, + ) + page_token = proto.Field( + proto.STRING, + number=3, + ) + page_size = proto.Field( + proto.INT32, + number=4, + ) class ListDataSourcesResponse(proto.Message): @@ -289,8 +405,15 @@ class ListDataSourcesResponse(proto.Message): def raw_page(self): return self - data_sources = proto.RepeatedField(proto.MESSAGE, number=1, message="DataSource",) - next_page_token = proto.Field(proto.STRING, number=2,) + data_sources = proto.RepeatedField( + proto.MESSAGE, + number=1, + message="DataSource", + ) + next_page_token = proto.Field( + proto.STRING, + number=2, + ) class CreateTransferConfigRequest(proto.Message): @@ -348,13 +471,27 @@ class CreateTransferConfigRequest(proto.Message): permissions to act as this service account. """ - parent = proto.Field(proto.STRING, number=1,) + parent = proto.Field( + proto.STRING, + number=1, + ) transfer_config = proto.Field( - proto.MESSAGE, number=2, message=transfer.TransferConfig, + proto.MESSAGE, + number=2, + message=transfer.TransferConfig, + ) + authorization_code = proto.Field( + proto.STRING, + number=3, + ) + version_info = proto.Field( + proto.STRING, + number=5, + ) + service_account_name = proto.Field( + proto.STRING, + number=6, ) - authorization_code = proto.Field(proto.STRING, number=3,) - version_info = proto.Field(proto.STRING, number=5,) - service_account_name = proto.Field(proto.STRING, number=6,) class UpdateTransferConfigRequest(proto.Message): @@ -406,14 +543,27 @@ class UpdateTransferConfigRequest(proto.Message): """ transfer_config = proto.Field( - proto.MESSAGE, number=1, message=transfer.TransferConfig, + proto.MESSAGE, + number=1, + message=transfer.TransferConfig, + ) + authorization_code = proto.Field( + proto.STRING, + number=3, ) - authorization_code = proto.Field(proto.STRING, number=3,) update_mask = proto.Field( - proto.MESSAGE, number=4, message=field_mask_pb2.FieldMask, + proto.MESSAGE, + number=4, + message=field_mask_pb2.FieldMask, + ) + version_info = proto.Field( + proto.STRING, + number=5, + ) + service_account_name = proto.Field( + proto.STRING, + number=6, ) - version_info = proto.Field(proto.STRING, number=5,) - service_account_name = proto.Field(proto.STRING, number=6,) class GetTransferConfigRequest(proto.Message): @@ -427,7 +577,10 @@ class GetTransferConfigRequest(proto.Message): ``projects/{project_id}/locations/{location_id}/transferConfigs/{config_id}`` """ - name = proto.Field(proto.STRING, number=1,) + name = proto.Field( + proto.STRING, + number=1, + ) class DeleteTransferConfigRequest(proto.Message): @@ -442,7 +595,10 @@ class DeleteTransferConfigRequest(proto.Message): ``projects/{project_id}/locations/{location_id}/transferConfigs/{config_id}`` """ - name = proto.Field(proto.STRING, number=1,) + name = proto.Field( + proto.STRING, + number=1, + ) class GetTransferRunRequest(proto.Message): @@ -457,7 +613,10 @@ class GetTransferRunRequest(proto.Message): ``projects/{project_id}/locations/{location_id}/transferConfigs/{config_id}/runs/{run_id}`` """ - name = proto.Field(proto.STRING, number=1,) + name = proto.Field( + proto.STRING, + number=1, + ) class DeleteTransferRunRequest(proto.Message): @@ -472,7 +631,10 @@ class DeleteTransferRunRequest(proto.Message): ``projects/{project_id}/locations/{location_id}/transferConfigs/{config_id}/runs/{run_id}`` """ - name = proto.Field(proto.STRING, number=1,) + name = proto.Field( + proto.STRING, + number=1, + ) class ListTransferConfigsRequest(proto.Message): @@ -498,10 +660,22 @@ class ListTransferConfigsRequest(proto.Message): maximum value of 1000 results. """ - parent = proto.Field(proto.STRING, number=1,) - data_source_ids = proto.RepeatedField(proto.STRING, number=2,) - page_token = proto.Field(proto.STRING, number=3,) - page_size = proto.Field(proto.INT32, number=4,) + parent = proto.Field( + proto.STRING, + number=1, + ) + data_source_ids = proto.RepeatedField( + proto.STRING, + number=2, + ) + page_token = proto.Field( + proto.STRING, + number=3, + ) + page_size = proto.Field( + proto.INT32, + number=4, + ) class ListTransferConfigsResponse(proto.Message): @@ -523,9 +697,14 @@ def raw_page(self): return self transfer_configs = proto.RepeatedField( - proto.MESSAGE, number=1, message=transfer.TransferConfig, + proto.MESSAGE, + number=1, + message=transfer.TransferConfig, + ) + next_page_token = proto.Field( + proto.STRING, + number=2, ) - next_page_token = proto.Field(proto.STRING, number=2,) class ListTransferRunsRequest(proto.Message): @@ -560,11 +739,28 @@ class RunAttempt(proto.Enum): RUN_ATTEMPT_UNSPECIFIED = 0 LATEST = 1 - parent = proto.Field(proto.STRING, number=1,) - states = proto.RepeatedField(proto.ENUM, number=2, enum=transfer.TransferState,) - page_token = proto.Field(proto.STRING, number=3,) - page_size = proto.Field(proto.INT32, number=4,) - run_attempt = proto.Field(proto.ENUM, number=5, enum=RunAttempt,) + parent = proto.Field( + proto.STRING, + number=1, + ) + states = proto.RepeatedField( + proto.ENUM, + number=2, + enum=transfer.TransferState, + ) + page_token = proto.Field( + proto.STRING, + number=3, + ) + page_size = proto.Field( + proto.INT32, + number=4, + ) + run_attempt = proto.Field( + proto.ENUM, + number=5, + enum=RunAttempt, + ) class ListTransferRunsResponse(proto.Message): @@ -586,9 +782,14 @@ def raw_page(self): return self transfer_runs = proto.RepeatedField( - proto.MESSAGE, number=1, message=transfer.TransferRun, + proto.MESSAGE, + number=1, + message=transfer.TransferRun, + ) + next_page_token = proto.Field( + proto.STRING, + number=2, ) - next_page_token = proto.Field(proto.STRING, number=2,) class ListTransferLogsRequest(proto.Message): @@ -616,11 +817,22 @@ class ListTransferLogsRequest(proto.Message): INFO, WARNING and ERROR messages are returned. """ - parent = proto.Field(proto.STRING, number=1,) - page_token = proto.Field(proto.STRING, number=4,) - page_size = proto.Field(proto.INT32, number=5,) + parent = proto.Field( + proto.STRING, + number=1, + ) + page_token = proto.Field( + proto.STRING, + number=4, + ) + page_size = proto.Field( + proto.INT32, + number=5, + ) message_types = proto.RepeatedField( - proto.ENUM, number=6, enum=transfer.TransferMessage.MessageSeverity, + proto.ENUM, + number=6, + enum=transfer.TransferMessage.MessageSeverity, ) @@ -643,9 +855,14 @@ def raw_page(self): return self transfer_messages = proto.RepeatedField( - proto.MESSAGE, number=1, message=transfer.TransferMessage, + proto.MESSAGE, + number=1, + message=transfer.TransferMessage, + ) + next_page_token = proto.Field( + proto.STRING, + number=2, ) - next_page_token = proto.Field(proto.STRING, number=2,) class CheckValidCredsRequest(proto.Message): @@ -664,7 +881,10 @@ class CheckValidCredsRequest(proto.Message): ``projects/{project_id}/locations/{location_id}/dataSources/{data_source_id}``. """ - name = proto.Field(proto.STRING, number=1,) + name = proto.Field( + proto.STRING, + number=1, + ) class CheckValidCredsResponse(proto.Message): @@ -676,7 +896,10 @@ class CheckValidCredsResponse(proto.Message): If set to ``true``, the credentials exist and are valid. """ - has_valid_creds = proto.Field(proto.BOOL, number=1,) + has_valid_creds = proto.Field( + proto.BOOL, + number=1, + ) class ScheduleTransferRunsRequest(proto.Message): @@ -695,9 +918,20 @@ class ScheduleTransferRunsRequest(proto.Message): example, ``"2017-05-30T00:00:00+00:00"``. """ - parent = proto.Field(proto.STRING, number=1,) - start_time = proto.Field(proto.MESSAGE, number=2, message=timestamp_pb2.Timestamp,) - end_time = proto.Field(proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp,) + parent = proto.Field( + proto.STRING, + number=1, + ) + start_time = proto.Field( + proto.MESSAGE, + number=2, + message=timestamp_pb2.Timestamp, + ) + end_time = proto.Field( + proto.MESSAGE, + number=3, + message=timestamp_pb2.Timestamp, + ) class ScheduleTransferRunsResponse(proto.Message): @@ -708,7 +942,11 @@ class ScheduleTransferRunsResponse(proto.Message): The transfer runs that were scheduled. """ - runs = proto.RepeatedField(proto.MESSAGE, number=1, message=transfer.TransferRun,) + runs = proto.RepeatedField( + proto.MESSAGE, + number=1, + message=transfer.TransferRun, + ) class StartManualTransferRunsRequest(proto.Message): @@ -759,18 +997,31 @@ class TimeRange(proto.Message): """ start_time = proto.Field( - proto.MESSAGE, number=1, message=timestamp_pb2.Timestamp, + proto.MESSAGE, + number=1, + message=timestamp_pb2.Timestamp, ) end_time = proto.Field( - proto.MESSAGE, number=2, message=timestamp_pb2.Timestamp, + proto.MESSAGE, + number=2, + message=timestamp_pb2.Timestamp, ) - parent = proto.Field(proto.STRING, number=1,) + parent = proto.Field( + proto.STRING, + number=1, + ) requested_time_range = proto.Field( - proto.MESSAGE, number=3, oneof="time", message=TimeRange, + proto.MESSAGE, + number=3, + oneof="time", + message=TimeRange, ) requested_run_time = proto.Field( - proto.MESSAGE, number=4, oneof="time", message=timestamp_pb2.Timestamp, + proto.MESSAGE, + number=4, + oneof="time", + message=timestamp_pb2.Timestamp, ) @@ -782,7 +1033,11 @@ class StartManualTransferRunsResponse(proto.Message): The transfer runs that were created. """ - runs = proto.RepeatedField(proto.MESSAGE, number=1, message=transfer.TransferRun,) + runs = proto.RepeatedField( + proto.MESSAGE, + number=1, + message=transfer.TransferRun, + ) class EnrollDataSourcesRequest(proto.Message): @@ -798,8 +1053,14 @@ class EnrollDataSourcesRequest(proto.Message): required to provide at least one data source id. """ - name = proto.Field(proto.STRING, number=1,) - data_source_ids = proto.RepeatedField(proto.STRING, number=2,) + name = proto.Field( + proto.STRING, + number=1, + ) + data_source_ids = proto.RepeatedField( + proto.STRING, + number=2, + ) __all__ = tuple(sorted(__protobuf__.manifest)) diff --git a/google/cloud/bigquery_datatransfer_v1/types/transfer.py b/google/cloud/bigquery_datatransfer_v1/types/transfer.py index 2f4975c7..f18f71e7 100644 --- a/google/cloud/bigquery_datatransfer_v1/types/transfer.py +++ b/google/cloud/bigquery_datatransfer_v1/types/transfer.py @@ -13,12 +13,10 @@ # See the License for the specific language governing permissions and # limitations under the License. # -import proto # type: ignore - from google.protobuf import struct_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore from google.rpc import status_pb2 # type: ignore - +import proto # type: ignore __protobuf__ = proto.module( package="google.cloud.bigquery.datatransfer.v1", @@ -63,7 +61,10 @@ class EmailPreferences(proto.Message): transfer run failures. """ - enable_failure_email = proto.Field(proto.BOOL, number=1,) + enable_failure_email = proto.Field( + proto.BOOL, + number=1, + ) class ScheduleOptions(proto.Message): @@ -94,9 +95,20 @@ class ScheduleOptions(proto.Message): option. """ - disable_auto_scheduling = proto.Field(proto.BOOL, number=3,) - start_time = proto.Field(proto.MESSAGE, number=1, message=timestamp_pb2.Timestamp,) - end_time = proto.Field(proto.MESSAGE, number=2, message=timestamp_pb2.Timestamp,) + disable_auto_scheduling = proto.Field( + proto.BOOL, + number=3, + ) + start_time = proto.Field( + proto.MESSAGE, + number=1, + message=timestamp_pb2.Timestamp, + ) + end_time = proto.Field( + proto.MESSAGE, + number=2, + message=timestamp_pb2.Timestamp, + ) class UserInfo(proto.Message): @@ -109,7 +121,11 @@ class UserInfo(proto.Message): This field is a member of `oneof`_ ``_email``. """ - email = proto.Field(proto.STRING, number=1, optional=True,) + email = proto.Field( + proto.STRING, + number=1, + optional=True, + ) class TransferConfig(proto.Message): @@ -210,28 +226,82 @@ class TransferConfig(proto.Message): This field is a member of `oneof`_ ``_owner_info``. """ - name = proto.Field(proto.STRING, number=1,) - destination_dataset_id = proto.Field(proto.STRING, number=2, oneof="destination",) - display_name = proto.Field(proto.STRING, number=3,) - data_source_id = proto.Field(proto.STRING, number=5,) - params = proto.Field(proto.MESSAGE, number=9, message=struct_pb2.Struct,) - schedule = proto.Field(proto.STRING, number=7,) - schedule_options = proto.Field(proto.MESSAGE, number=24, message="ScheduleOptions",) - data_refresh_window_days = proto.Field(proto.INT32, number=12,) - disabled = proto.Field(proto.BOOL, number=13,) - update_time = proto.Field(proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp,) + name = proto.Field( + proto.STRING, + number=1, + ) + destination_dataset_id = proto.Field( + proto.STRING, + number=2, + oneof="destination", + ) + display_name = proto.Field( + proto.STRING, + number=3, + ) + data_source_id = proto.Field( + proto.STRING, + number=5, + ) + params = proto.Field( + proto.MESSAGE, + number=9, + message=struct_pb2.Struct, + ) + schedule = proto.Field( + proto.STRING, + number=7, + ) + schedule_options = proto.Field( + proto.MESSAGE, + number=24, + message="ScheduleOptions", + ) + data_refresh_window_days = proto.Field( + proto.INT32, + number=12, + ) + disabled = proto.Field( + proto.BOOL, + number=13, + ) + update_time = proto.Field( + proto.MESSAGE, + number=4, + message=timestamp_pb2.Timestamp, + ) next_run_time = proto.Field( - proto.MESSAGE, number=8, message=timestamp_pb2.Timestamp, + proto.MESSAGE, + number=8, + message=timestamp_pb2.Timestamp, + ) + state = proto.Field( + proto.ENUM, + number=10, + enum="TransferState", + ) + user_id = proto.Field( + proto.INT64, + number=11, + ) + dataset_region = proto.Field( + proto.STRING, + number=14, + ) + notification_pubsub_topic = proto.Field( + proto.STRING, + number=15, ) - state = proto.Field(proto.ENUM, number=10, enum="TransferState",) - user_id = proto.Field(proto.INT64, number=11,) - dataset_region = proto.Field(proto.STRING, number=14,) - notification_pubsub_topic = proto.Field(proto.STRING, number=15,) email_preferences = proto.Field( - proto.MESSAGE, number=18, message="EmailPreferences", + proto.MESSAGE, + number=18, + message="EmailPreferences", ) owner_info = proto.Field( - proto.MESSAGE, number=27, optional=True, message="UserInfo", + proto.MESSAGE, + number=27, + optional=True, + message="UserInfo", ) @@ -303,24 +373,75 @@ class TransferRun(proto.Message): this run was derived from. """ - name = proto.Field(proto.STRING, number=1,) + name = proto.Field( + proto.STRING, + number=1, + ) schedule_time = proto.Field( - proto.MESSAGE, number=3, message=timestamp_pb2.Timestamp, - ) - run_time = proto.Field(proto.MESSAGE, number=10, message=timestamp_pb2.Timestamp,) - error_status = proto.Field(proto.MESSAGE, number=21, message=status_pb2.Status,) - start_time = proto.Field(proto.MESSAGE, number=4, message=timestamp_pb2.Timestamp,) - end_time = proto.Field(proto.MESSAGE, number=5, message=timestamp_pb2.Timestamp,) - update_time = proto.Field(proto.MESSAGE, number=6, message=timestamp_pb2.Timestamp,) - params = proto.Field(proto.MESSAGE, number=9, message=struct_pb2.Struct,) - destination_dataset_id = proto.Field(proto.STRING, number=2, oneof="destination",) - data_source_id = proto.Field(proto.STRING, number=7,) - state = proto.Field(proto.ENUM, number=8, enum="TransferState",) - user_id = proto.Field(proto.INT64, number=11,) - schedule = proto.Field(proto.STRING, number=12,) - notification_pubsub_topic = proto.Field(proto.STRING, number=23,) + proto.MESSAGE, + number=3, + message=timestamp_pb2.Timestamp, + ) + run_time = proto.Field( + proto.MESSAGE, + number=10, + message=timestamp_pb2.Timestamp, + ) + error_status = proto.Field( + proto.MESSAGE, + number=21, + message=status_pb2.Status, + ) + start_time = proto.Field( + proto.MESSAGE, + number=4, + message=timestamp_pb2.Timestamp, + ) + end_time = proto.Field( + proto.MESSAGE, + number=5, + message=timestamp_pb2.Timestamp, + ) + update_time = proto.Field( + proto.MESSAGE, + number=6, + message=timestamp_pb2.Timestamp, + ) + params = proto.Field( + proto.MESSAGE, + number=9, + message=struct_pb2.Struct, + ) + destination_dataset_id = proto.Field( + proto.STRING, + number=2, + oneof="destination", + ) + data_source_id = proto.Field( + proto.STRING, + number=7, + ) + state = proto.Field( + proto.ENUM, + number=8, + enum="TransferState", + ) + user_id = proto.Field( + proto.INT64, + number=11, + ) + schedule = proto.Field( + proto.STRING, + number=12, + ) + notification_pubsub_topic = proto.Field( + proto.STRING, + number=23, + ) email_preferences = proto.Field( - proto.MESSAGE, number=25, message="EmailPreferences", + proto.MESSAGE, + number=25, + message="EmailPreferences", ) @@ -345,10 +466,19 @@ class MessageSeverity(proto.Enum): ERROR = 3 message_time = proto.Field( - proto.MESSAGE, number=1, message=timestamp_pb2.Timestamp, + proto.MESSAGE, + number=1, + message=timestamp_pb2.Timestamp, + ) + severity = proto.Field( + proto.ENUM, + number=2, + enum=MessageSeverity, + ) + message_text = proto.Field( + proto.STRING, + number=3, ) - severity = proto.Field(proto.ENUM, number=2, enum=MessageSeverity,) - message_text = proto.Field(proto.STRING, number=3,) __all__ = tuple(sorted(__protobuf__.manifest)) diff --git a/noxfile.py b/noxfile.py index 2a2001c4..7c1742d2 100644 --- a/noxfile.py +++ b/noxfile.py @@ -17,19 +17,45 @@ # Generated by synthtool. DO NOT EDIT! from __future__ import absolute_import + import os import pathlib import shutil +import warnings import nox - -BLACK_VERSION = "black==19.10b0" -BLACK_PATHS = ["docs", "google", "tests", "noxfile.py", "setup.py"] +BLACK_VERSION = "black==22.3.0" +ISORT_VERSION = "isort==5.10.1" +LINT_PATHS = ["docs", "google", "tests", "noxfile.py", "setup.py"] DEFAULT_PYTHON_VERSION = "3.8" -SYSTEM_TEST_PYTHON_VERSIONS = ["3.8"] + UNIT_TEST_PYTHON_VERSIONS = ["3.6", "3.7", "3.8", "3.9", "3.10"] +UNIT_TEST_STANDARD_DEPENDENCIES = [ + "mock", + "asyncmock", + "pytest", + "pytest-cov", + "pytest-asyncio", +] +UNIT_TEST_EXTERNAL_DEPENDENCIES = [] +UNIT_TEST_LOCAL_DEPENDENCIES = [] +UNIT_TEST_DEPENDENCIES = [] +UNIT_TEST_EXTRAS = [] +UNIT_TEST_EXTRAS_BY_PYTHON = {} + +SYSTEM_TEST_PYTHON_VERSIONS = ["3.8"] +SYSTEM_TEST_STANDARD_DEPENDENCIES = [ + "mock", + "pytest", + "google-cloud-testutils", +] +SYSTEM_TEST_EXTERNAL_DEPENDENCIES = [] +SYSTEM_TEST_LOCAL_DEPENDENCIES = [] +SYSTEM_TEST_DEPENDENCIES = [] +SYSTEM_TEST_EXTRAS = [] +SYSTEM_TEST_EXTRAS_BY_PYTHON = {} CURRENT_DIRECTORY = pathlib.Path(__file__).parent.absolute() @@ -57,7 +83,9 @@ def lint(session): """ session.install("flake8", BLACK_VERSION) session.run( - "black", "--check", *BLACK_PATHS, + "black", + "--check", + *LINT_PATHS, ) session.run("flake8", "google", "tests") @@ -67,7 +95,28 @@ def blacken(session): """Run black. Format code to uniform standard.""" session.install(BLACK_VERSION) session.run( - "black", *BLACK_PATHS, + "black", + *LINT_PATHS, + ) + + +@nox.session(python=DEFAULT_PYTHON_VERSION) +def format(session): + """ + Run isort to sort imports. Then run black + to format code to uniform standard. + """ + session.install(BLACK_VERSION, ISORT_VERSION) + # Use the --fss option to sort imports using strict alphabetical order. + # See https://pycqa.github.io/isort/docs/configuration/options.html#force-sort-within-sections + session.run( + "isort", + "--fss", + *LINT_PATHS, + ) + session.run( + "black", + *LINT_PATHS, ) @@ -78,23 +127,41 @@ def lint_setup_py(session): session.run("python", "setup.py", "check", "--restructuredtext", "--strict") +def install_unittest_dependencies(session, *constraints): + standard_deps = UNIT_TEST_STANDARD_DEPENDENCIES + UNIT_TEST_DEPENDENCIES + session.install(*standard_deps, *constraints) + + if UNIT_TEST_EXTERNAL_DEPENDENCIES: + warnings.warn( + "'unit_test_external_dependencies' is deprecated. Instead, please " + "use 'unit_test_dependencies' or 'unit_test_local_dependencies'.", + DeprecationWarning, + ) + session.install(*UNIT_TEST_EXTERNAL_DEPENDENCIES, *constraints) + + if UNIT_TEST_LOCAL_DEPENDENCIES: + session.install(*UNIT_TEST_LOCAL_DEPENDENCIES, *constraints) + + if UNIT_TEST_EXTRAS_BY_PYTHON: + extras = UNIT_TEST_EXTRAS_BY_PYTHON.get(session.python, []) + elif UNIT_TEST_EXTRAS: + extras = UNIT_TEST_EXTRAS + else: + extras = [] + + if extras: + session.install("-e", f".[{','.join(extras)}]", *constraints) + else: + session.install("-e", ".", *constraints) + + def default(session): # Install all test dependencies, then install this package in-place. constraints_path = str( CURRENT_DIRECTORY / "testing" / f"constraints-{session.python}.txt" ) - session.install( - "mock", - "asyncmock", - "pytest", - "pytest-cov", - "pytest-asyncio", - "-c", - constraints_path, - ) - - session.install("-e", ".", "-c", constraints_path) + install_unittest_dependencies(session, "-c", constraints_path) # Run py.test against the unit tests. session.run( @@ -118,6 +185,35 @@ def unit(session): default(session) +def install_systemtest_dependencies(session, *constraints): + + # Use pre-release gRPC for system tests. + session.install("--pre", "grpcio") + + session.install(*SYSTEM_TEST_STANDARD_DEPENDENCIES, *constraints) + + if SYSTEM_TEST_EXTERNAL_DEPENDENCIES: + session.install(*SYSTEM_TEST_EXTERNAL_DEPENDENCIES, *constraints) + + if SYSTEM_TEST_LOCAL_DEPENDENCIES: + session.install("-e", *SYSTEM_TEST_LOCAL_DEPENDENCIES, *constraints) + + if SYSTEM_TEST_DEPENDENCIES: + session.install("-e", *SYSTEM_TEST_DEPENDENCIES, *constraints) + + if SYSTEM_TEST_EXTRAS_BY_PYTHON: + extras = SYSTEM_TEST_EXTRAS_BY_PYTHON.get(session.python, []) + elif SYSTEM_TEST_EXTRAS: + extras = SYSTEM_TEST_EXTRAS + else: + extras = [] + + if extras: + session.install("-e", f".[{','.join(extras)}]", *constraints) + else: + session.install("-e", ".", *constraints) + + @nox.session(python=SYSTEM_TEST_PYTHON_VERSIONS) def system(session): """Run the system test suite.""" @@ -140,13 +236,7 @@ def system(session): if not system_test_exists and not system_test_folder_exists: session.skip("System tests were not found") - # Use pre-release gRPC for system tests. - session.install("--pre", "grpcio") - - # Install all test dependencies, then install this package into the - # virtualenv's dist-packages. - session.install("mock", "pytest", "google-cloud-testutils", "-c", constraints_path) - session.install("-e", ".", "-c", constraints_path) + install_systemtest_dependencies(session, "-c", constraints_path) # Run py.test against the system tests. if system_test_exists: diff --git a/samples/generated_samples/snippet_metadata_bigquery datatransfer_v1.json b/samples/generated_samples/snippet_metadata_bigquery datatransfer_v1.json index 45efb07d..d94c3641 100644 --- a/samples/generated_samples/snippet_metadata_bigquery datatransfer_v1.json +++ b/samples/generated_samples/snippet_metadata_bigquery datatransfer_v1.json @@ -1,16 +1,61 @@ { + "clientLibrary": { + "apis": [ + { + "id": "google.cloud.bigquery.datatransfer.v1", + "version": "v1" + } + ], + "language": "PYTHON", + "name": "google-cloud-bigquery-datatransfer" + }, "snippets": [ { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.check_valid_creds", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.CheckValidCreds", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "CheckValidCreds" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.CheckValidCredsRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.CheckValidCredsResponse", + "shortName": "check_valid_creds" }, + "description": "Sample for CheckValidCreds", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_check_valid_creds_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_CheckValidCreds_async", "segments": [ { @@ -43,18 +88,54 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_check_valid_creds_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.check_valid_creds", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.CheckValidCreds", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "CheckValidCreds" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.CheckValidCredsRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.CheckValidCredsResponse", + "shortName": "check_valid_creds" }, + "description": "Sample for CheckValidCreds", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_check_valid_creds_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_CheckValidCreds_sync", "segments": [ { @@ -87,19 +168,59 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_check_valid_creds_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.create_transfer_config", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.CreateTransferConfig", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "CreateTransferConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.CreateTransferConfigRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "transfer_config", + "type": "google.cloud.bigquery_datatransfer_v1.types.TransferConfig" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.TransferConfig", + "shortName": "create_transfer_config" }, + "description": "Sample for CreateTransferConfig", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_create_transfer_config_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_CreateTransferConfig_async", "segments": [ { @@ -132,18 +253,58 @@ "start": 46, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_create_transfer_config_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.create_transfer_config", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.CreateTransferConfig", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "CreateTransferConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.CreateTransferConfigRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "transfer_config", + "type": "google.cloud.bigquery_datatransfer_v1.types.TransferConfig" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.TransferConfig", + "shortName": "create_transfer_config" }, + "description": "Sample for CreateTransferConfig", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_create_transfer_config_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_CreateTransferConfig_sync", "segments": [ { @@ -176,19 +337,54 @@ "start": 46, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_create_transfer_config_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.delete_transfer_config", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.DeleteTransferConfig", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "DeleteTransferConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.DeleteTransferConfigRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "shortName": "delete_transfer_config" }, + "description": "Sample for DeleteTransferConfig", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_delete_transfer_config_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_DeleteTransferConfig_async", "segments": [ { @@ -219,18 +415,53 @@ "end": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_delete_transfer_config_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.delete_transfer_config", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.DeleteTransferConfig", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "DeleteTransferConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.DeleteTransferConfigRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "shortName": "delete_transfer_config" }, + "description": "Sample for DeleteTransferConfig", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_delete_transfer_config_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_DeleteTransferConfig_sync", "segments": [ { @@ -261,19 +492,54 @@ "end": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_delete_transfer_config_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.delete_transfer_run", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.DeleteTransferRun", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "DeleteTransferRun" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.DeleteTransferRunRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "shortName": "delete_transfer_run" }, + "description": "Sample for DeleteTransferRun", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_delete_transfer_run_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_DeleteTransferRun_async", "segments": [ { @@ -304,18 +570,53 @@ "end": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_delete_transfer_run_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.delete_transfer_run", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.DeleteTransferRun", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "DeleteTransferRun" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.DeleteTransferRunRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "shortName": "delete_transfer_run" }, + "description": "Sample for DeleteTransferRun", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_delete_transfer_run_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_DeleteTransferRun_sync", "segments": [ { @@ -346,19 +647,50 @@ "end": 43, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_delete_transfer_run_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.enroll_data_sources", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.EnrollDataSources", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "EnrollDataSources" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.EnrollDataSourcesRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "shortName": "enroll_data_sources" }, + "description": "Sample for EnrollDataSources", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_enroll_data_sources_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_EnrollDataSources_async", "segments": [ { @@ -389,18 +721,49 @@ "end": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_enroll_data_sources_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.enroll_data_sources", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.EnrollDataSources", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "EnrollDataSources" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.EnrollDataSourcesRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "shortName": "enroll_data_sources" }, + "description": "Sample for EnrollDataSources", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_enroll_data_sources_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_EnrollDataSources_sync", "segments": [ { @@ -431,19 +794,55 @@ "end": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_enroll_data_sources_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.get_data_source", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.GetDataSource", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "GetDataSource" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.GetDataSourceRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.DataSource", + "shortName": "get_data_source" }, + "description": "Sample for GetDataSource", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_get_data_source_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_GetDataSource_async", "segments": [ { @@ -476,18 +875,54 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_get_data_source_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.get_data_source", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.GetDataSource", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "GetDataSource" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.GetDataSourceRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.DataSource", + "shortName": "get_data_source" }, + "description": "Sample for GetDataSource", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_get_data_source_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_GetDataSource_sync", "segments": [ { @@ -520,19 +955,55 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_get_data_source_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.get_transfer_config", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.GetTransferConfig", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "GetTransferConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.GetTransferConfigRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.TransferConfig", + "shortName": "get_transfer_config" }, + "description": "Sample for GetTransferConfig", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_get_transfer_config_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_GetTransferConfig_async", "segments": [ { @@ -565,18 +1036,54 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_get_transfer_config_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.get_transfer_config", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.GetTransferConfig", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "GetTransferConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.GetTransferConfigRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.TransferConfig", + "shortName": "get_transfer_config" }, + "description": "Sample for GetTransferConfig", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_get_transfer_config_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_GetTransferConfig_sync", "segments": [ { @@ -609,19 +1116,55 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_get_transfer_config_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.get_transfer_run", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.GetTransferRun", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "GetTransferRun" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.GetTransferRunRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.TransferRun", + "shortName": "get_transfer_run" }, + "description": "Sample for GetTransferRun", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_get_transfer_run_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_GetTransferRun_async", "segments": [ { @@ -654,18 +1197,54 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_get_transfer_run_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.get_transfer_run", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.GetTransferRun", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "GetTransferRun" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.GetTransferRunRequest" + }, + { + "name": "name", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.TransferRun", + "shortName": "get_transfer_run" }, + "description": "Sample for GetTransferRun", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_get_transfer_run_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_GetTransferRun_sync", "segments": [ { @@ -698,19 +1277,55 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_get_transfer_run_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.list_data_sources", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.ListDataSources", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "ListDataSources" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.ListDataSourcesRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.services.data_transfer_service.pagers.ListDataSourcesAsyncPager", + "shortName": "list_data_sources" }, + "description": "Sample for ListDataSources", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_list_data_sources_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_ListDataSources_async", "segments": [ { @@ -743,18 +1358,54 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_list_data_sources_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.list_data_sources", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.ListDataSources", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "ListDataSources" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.ListDataSourcesRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.services.data_transfer_service.pagers.ListDataSourcesPager", + "shortName": "list_data_sources" }, + "description": "Sample for ListDataSources", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_list_data_sources_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_ListDataSources_sync", "segments": [ { @@ -787,19 +1438,55 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_list_data_sources_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.list_transfer_configs", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.ListTransferConfigs", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "ListTransferConfigs" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.ListTransferConfigsRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.services.data_transfer_service.pagers.ListTransferConfigsAsyncPager", + "shortName": "list_transfer_configs" }, + "description": "Sample for ListTransferConfigs", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_configs_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_ListTransferConfigs_async", "segments": [ { @@ -832,18 +1519,54 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_configs_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.list_transfer_configs", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.ListTransferConfigs", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "ListTransferConfigs" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.ListTransferConfigsRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.services.data_transfer_service.pagers.ListTransferConfigsPager", + "shortName": "list_transfer_configs" }, + "description": "Sample for ListTransferConfigs", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_configs_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_ListTransferConfigs_sync", "segments": [ { @@ -876,19 +1599,55 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_configs_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.list_transfer_logs", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.ListTransferLogs", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "ListTransferLogs" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.ListTransferLogsRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.services.data_transfer_service.pagers.ListTransferLogsAsyncPager", + "shortName": "list_transfer_logs" }, + "description": "Sample for ListTransferLogs", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_logs_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_ListTransferLogs_async", "segments": [ { @@ -921,18 +1680,54 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_logs_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.list_transfer_logs", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.ListTransferLogs", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "ListTransferLogs" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.ListTransferLogsRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.services.data_transfer_service.pagers.ListTransferLogsPager", + "shortName": "list_transfer_logs" }, + "description": "Sample for ListTransferLogs", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_logs_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_ListTransferLogs_sync", "segments": [ { @@ -965,19 +1760,55 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_logs_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.list_transfer_runs", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.ListTransferRuns", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "ListTransferRuns" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.ListTransferRunsRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.services.data_transfer_service.pagers.ListTransferRunsAsyncPager", + "shortName": "list_transfer_runs" }, + "description": "Sample for ListTransferRuns", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_runs_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_ListTransferRuns_async", "segments": [ { @@ -1010,18 +1841,54 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_runs_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.list_transfer_runs", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.ListTransferRuns", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "ListTransferRuns" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.ListTransferRunsRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.services.data_transfer_service.pagers.ListTransferRunsPager", + "shortName": "list_transfer_runs" }, + "description": "Sample for ListTransferRuns", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_runs_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_ListTransferRuns_sync", "segments": [ { @@ -1054,19 +1921,63 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_list_transfer_runs_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.schedule_transfer_runs", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.ScheduleTransferRuns", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "ScheduleTransferRuns" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.ScheduleTransferRunsRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "start_time", + "type": "google.protobuf.timestamp_pb2.Timestamp" + }, + { + "name": "end_time", + "type": "google.protobuf.timestamp_pb2.Timestamp" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.ScheduleTransferRunsResponse", + "shortName": "schedule_transfer_runs" }, + "description": "Sample for ScheduleTransferRuns", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_schedule_transfer_runs_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_ScheduleTransferRuns_async", "segments": [ { @@ -1099,18 +2010,62 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_schedule_transfer_runs_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.schedule_transfer_runs", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.ScheduleTransferRuns", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "ScheduleTransferRuns" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.ScheduleTransferRunsRequest" + }, + { + "name": "parent", + "type": "str" + }, + { + "name": "start_time", + "type": "google.protobuf.timestamp_pb2.Timestamp" + }, + { + "name": "end_time", + "type": "google.protobuf.timestamp_pb2.Timestamp" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.ScheduleTransferRunsResponse", + "shortName": "schedule_transfer_runs" }, + "description": "Sample for ScheduleTransferRuns", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_schedule_transfer_runs_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_ScheduleTransferRuns_sync", "segments": [ { @@ -1143,19 +2098,51 @@ "start": 42, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_schedule_transfer_runs_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.start_manual_transfer_runs", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.StartManualTransferRuns", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "StartManualTransferRuns" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.StartManualTransferRunsRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.StartManualTransferRunsResponse", + "shortName": "start_manual_transfer_runs" }, + "description": "Sample for StartManualTransferRuns", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_start_manual_transfer_runs_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_StartManualTransferRuns_async", "segments": [ { @@ -1188,18 +2175,50 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_start_manual_transfer_runs_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.start_manual_transfer_runs", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.StartManualTransferRuns", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "StartManualTransferRuns" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.StartManualTransferRunsRequest" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.StartManualTransferRunsResponse", + "shortName": "start_manual_transfer_runs" }, + "description": "Sample for StartManualTransferRuns", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_start_manual_transfer_runs_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_StartManualTransferRuns_sync", "segments": [ { @@ -1232,19 +2251,59 @@ "start": 41, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_start_manual_transfer_runs_sync.py" }, { + "canonical": true, "clientMethod": { "async": true, + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient", + "shortName": "DataTransferServiceAsyncClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceAsyncClient.update_transfer_config", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.UpdateTransferConfig", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "UpdateTransferConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.UpdateTransferConfigRequest" + }, + { + "name": "transfer_config", + "type": "google.cloud.bigquery_datatransfer_v1.types.TransferConfig" + }, + { + "name": "update_mask", + "type": "google.protobuf.field_mask_pb2.FieldMask" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.TransferConfig", + "shortName": "update_transfer_config" }, + "description": "Sample for UpdateTransferConfig", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_update_transfer_config_async.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_UpdateTransferConfig_async", "segments": [ { @@ -1277,18 +2336,58 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_update_transfer_config_async.py" }, { + "canonical": true, "clientMethod": { + "client": { + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient", + "shortName": "DataTransferServiceClient" + }, + "fullName": "google.cloud.bigquery_datatransfer_v1.DataTransferServiceClient.update_transfer_config", "method": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService.UpdateTransferConfig", "service": { + "fullName": "google.cloud.bigquery.datatransfer.v1.DataTransferService", "shortName": "DataTransferService" }, "shortName": "UpdateTransferConfig" - } + }, + "parameters": [ + { + "name": "request", + "type": "google.cloud.bigquery_datatransfer_v1.types.UpdateTransferConfigRequest" + }, + { + "name": "transfer_config", + "type": "google.cloud.bigquery_datatransfer_v1.types.TransferConfig" + }, + { + "name": "update_mask", + "type": "google.protobuf.field_mask_pb2.FieldMask" + }, + { + "name": "retry", + "type": "google.api_core.retry.Retry" + }, + { + "name": "timeout", + "type": "float" + }, + { + "name": "metadata", + "type": "Sequence[Tuple[str, str]" + } + ], + "resultType": "google.cloud.bigquery_datatransfer_v1.types.TransferConfig", + "shortName": "update_transfer_config" }, + "description": "Sample for UpdateTransferConfig", "file": "bigquerydatatransfer_v1_generated_data_transfer_service_update_transfer_config_sync.py", + "language": "PYTHON", + "origin": "API_DEFINITION", "regionTag": "bigquerydatatransfer_v1_generated_DataTransferService_UpdateTransferConfig_sync", "segments": [ { @@ -1321,7 +2420,8 @@ "start": 45, "type": "RESPONSE_HANDLING" } - ] + ], + "title": "bigquerydatatransfer_v1_generated_data_transfer_service_update_transfer_config_sync.py" } ] } diff --git a/samples/snippets/conftest.py b/samples/snippets/conftest.py index f708ff48..1248a940 100644 --- a/samples/snippets/conftest.py +++ b/samples/snippets/conftest.py @@ -20,12 +20,9 @@ from google.api_core import client_options import google.api_core.exceptions import google.auth -from google.cloud import bigquery -from google.cloud import bigquery_datatransfer -from google.cloud import pubsub_v1 +from google.cloud import bigquery, bigquery_datatransfer, pubsub_v1 import pytest - RESOURCE_PREFIX = "python_bigquery_datatransfer_samples_snippets" RESOURCE_DATE_FORMAT = "%Y%m%d%H%M%S" RESOURCE_DATE_LENGTH = 4 + 2 + 2 + 2 + 2 + 2 diff --git a/samples/snippets/manage_transfer_configs.py b/samples/snippets/manage_transfer_configs.py index 5f775f10..6b4abd78 100644 --- a/samples/snippets/manage_transfer_configs.py +++ b/samples/snippets/manage_transfer_configs.py @@ -135,7 +135,9 @@ def schedule_backfill(override_values={}): ) response = transfer_client.schedule_transfer_runs( - parent=transfer_config_name, start_time=start_time, end_time=end_time, + parent=transfer_config_name, + start_time=start_time, + end_time=end_time, ) print("Started transfer runs:") diff --git a/samples/snippets/noxfile.py b/samples/snippets/noxfile.py index 85f5836d..3b3ffa5d 100644 --- a/samples/snippets/noxfile.py +++ b/samples/snippets/noxfile.py @@ -22,14 +22,14 @@ import nox - # WARNING - WARNING - WARNING - WARNING - WARNING # WARNING - WARNING - WARNING - WARNING - WARNING # DO NOT EDIT THIS FILE EVER! # WARNING - WARNING - WARNING - WARNING - WARNING # WARNING - WARNING - WARNING - WARNING - WARNING -BLACK_VERSION = "black==19.10b0" +BLACK_VERSION = "black==22.3.0" +ISORT_VERSION = "isort==5.10.1" # Copy `noxfile_config.py` to your directory and modify it instead. @@ -168,12 +168,33 @@ def lint(session: nox.sessions.Session) -> None: @nox.session def blacken(session: nox.sessions.Session) -> None: + """Run black. Format code to uniform standard.""" session.install(BLACK_VERSION) python_files = [path for path in os.listdir(".") if path.endswith(".py")] session.run("black", *python_files) +# +# format = isort + black +# + + +@nox.session +def format(session: nox.sessions.Session) -> None: + """ + Run isort to sort imports. Then run black + to format code to uniform standard. + """ + session.install(BLACK_VERSION, ISORT_VERSION) + python_files = [path for path in os.listdir(".") if path.endswith(".py")] + + # Use the --fss option to sort imports using strict alphabetical order. + # See https://pycqa.github.io/isort/docs/configuration/options.html#force-sort-within-sections + session.run("isort", "--fss", *python_files) + session.run("black", *python_files) + + # # Sample Tests # @@ -253,7 +274,7 @@ def py(session: nox.sessions.Session) -> None: def _get_repo_root() -> Optional[str]: - """ Returns the root folder of the project. """ + """Returns the root folder of the project.""" # Get root of this repository. Assume we don't have directories nested deeper than 10 items. p = Path(os.getcwd()) for i in range(10): diff --git a/samples/snippets/requirements-test.txt b/samples/snippets/requirements-test.txt index 6e81a536..a7d6e0cf 100644 --- a/samples/snippets/requirements-test.txt +++ b/samples/snippets/requirements-test.txt @@ -1,4 +1,4 @@ -google-cloud-bigquery==2.34.1 -google-cloud-pubsub==2.10.0 -pytest==7.0.1 +google-cloud-bigquery==3.1.0 +google-cloud-pubsub==2.12.1 +pytest==7.1.2 mock==4.0.3 diff --git a/samples/snippets/requirements.txt b/samples/snippets/requirements.txt index 6286e0dc..ef77f23f 100644 --- a/samples/snippets/requirements.txt +++ b/samples/snippets/requirements.txt @@ -1 +1 @@ -google-cloud-bigquery-datatransfer==3.6.0 +google-cloud-bigquery-datatransfer==3.6.1 diff --git a/samples/snippets/run_notification_test.py b/samples/snippets/run_notification_test.py index 4c41e689..02f24266 100644 --- a/samples/snippets/run_notification_test.py +++ b/samples/snippets/run_notification_test.py @@ -17,7 +17,8 @@ def test_run_notification(capsys, transfer_config_name, pubsub_topic): run_notification.run_notification( - transfer_config_name=transfer_config_name, pubsub_topic=pubsub_topic, + transfer_config_name=transfer_config_name, + pubsub_topic=pubsub_topic, ) out, _ = capsys.readouterr() assert "Updated config:" in out diff --git a/scripts/readme-gen/readme_gen.py b/scripts/readme-gen/readme_gen.py index d309d6e9..91b59676 100644 --- a/scripts/readme-gen/readme_gen.py +++ b/scripts/readme-gen/readme_gen.py @@ -28,7 +28,10 @@ jinja_env = jinja2.Environment( trim_blocks=True, loader=jinja2.FileSystemLoader( - os.path.abspath(os.path.join(os.path.dirname(__file__), 'templates')))) + os.path.abspath(os.path.join(os.path.dirname(__file__), "templates")) + ), + autoescape=True, +) README_TMPL = jinja_env.get_template('README.tmpl.rst') diff --git a/setup.py b/setup.py index 87c01371..a689335b 100644 --- a/setup.py +++ b/setup.py @@ -17,12 +17,11 @@ import setuptools - # Package metadata. name = "google-cloud-bigquery-datatransfer" description = "BigQuery Data Transfer API client library" -version = "3.6.1" +version = "3.6.2" # Should be one of: # 'Development Status :: 3 - Alpha' # 'Development Status :: 4 - Beta' @@ -33,8 +32,8 @@ # Until this issue is closed # https://github.com/googleapis/google-cloud-python/issues/10566 "google-api-core[grpc] >= 1.31.5, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.0", - "proto-plus >= 1.15.0", - "pytz >= 2021.1", + "proto-plus >= 1.15.0, <2.0.0dev", + "protobuf >= 3.19.0, <4.0.0dev", ) extras = {"libcst": "libcst >= 0.2.5"} diff --git a/testing/constraints-3.6.txt b/testing/constraints-3.6.txt index 9f829f6a..3b53664c 100644 --- a/testing/constraints-3.6.txt +++ b/testing/constraints-3.6.txt @@ -8,3 +8,4 @@ google-api-core==1.31.5 proto-plus==1.15.0 libcst==0.2.5 +protobuf==3.19.0 diff --git a/testing/constraints-3.7.txt b/testing/constraints-3.7.txt index e69de29b..3b53664c 100644 --- a/testing/constraints-3.7.txt +++ b/testing/constraints-3.7.txt @@ -0,0 +1,11 @@ +# This constraints file is used to check that lower bounds +# are correct in setup.py +# List *all* library dependencies and extras in this file. +# Pin the version to the lower bound. +# +# e.g., if setup.py has "foo >= 1.14.0, < 2.0.0dev", +# Then this file should have foo==1.14.0 +google-api-core==1.31.5 +proto-plus==1.15.0 +libcst==0.2.5 +protobuf==3.19.0 diff --git a/tests/system.py b/tests/system.py index bce6257e..03e24313 100644 --- a/tests/system.py +++ b/tests/system.py @@ -15,6 +15,7 @@ # limitations under the License. import os + import pytest from google.cloud import bigquery_datatransfer diff --git a/tests/unit/gapic/bigquery_datatransfer_v1/test_data_transfer_service.py b/tests/unit/gapic/bigquery_datatransfer_v1/test_data_transfer_service.py index b21d9678..98281f61 100644 --- a/tests/unit/gapic/bigquery_datatransfer_v1/test_data_transfer_service.py +++ b/tests/unit/gapic/bigquery_datatransfer_v1/test_data_transfer_service.py @@ -14,42 +14,40 @@ # limitations under the License. # import os -import mock -import grpc -from grpc.experimental import aio -import math -import pytest -from proto.marshal.rules.dates import DurationRule, TimestampRule +# try/except added for compatibility with python < 3.8 +try: + from unittest import mock + from unittest.mock import AsyncMock +except ImportError: + import mock +import math +from google.api_core import gapic_v1, grpc_helpers, grpc_helpers_async, path_template from google.api_core import client_options from google.api_core import exceptions as core_exceptions -from google.api_core import gapic_v1 -from google.api_core import grpc_helpers -from google.api_core import grpc_helpers_async -from google.api_core import path_template +import google.auth from google.auth import credentials as ga_credentials from google.auth.exceptions import MutualTLSChannelError -from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service import ( - DataTransferServiceAsyncClient, -) -from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service import ( - DataTransferServiceClient, -) -from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service import pagers -from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service import ( - transports, -) -from google.cloud.bigquery_datatransfer_v1.types import datatransfer -from google.cloud.bigquery_datatransfer_v1.types import transfer from google.oauth2 import service_account from google.protobuf import duration_pb2 # type: ignore from google.protobuf import field_mask_pb2 # type: ignore from google.protobuf import struct_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore from google.rpc import status_pb2 # type: ignore -import google.auth +import grpc +from grpc.experimental import aio +from proto.marshal.rules.dates import DurationRule, TimestampRule +import pytest + +from google.cloud.bigquery_datatransfer_v1.services.data_transfer_service import ( + DataTransferServiceAsyncClient, + DataTransferServiceClient, + pagers, + transports, +) +from google.cloud.bigquery_datatransfer_v1.types import datatransfer, transfer def client_cert_source_callback(): @@ -98,20 +96,26 @@ def test__get_default_mtls_endpoint(): @pytest.mark.parametrize( - "client_class", [DataTransferServiceClient, DataTransferServiceAsyncClient,] + "client_class,transport_name", + [ + (DataTransferServiceClient, "grpc"), + (DataTransferServiceAsyncClient, "grpc_asyncio"), + ], ) -def test_data_transfer_service_client_from_service_account_info(client_class): +def test_data_transfer_service_client_from_service_account_info( + client_class, transport_name +): creds = ga_credentials.AnonymousCredentials() with mock.patch.object( service_account.Credentials, "from_service_account_info" ) as factory: factory.return_value = creds info = {"valid": True} - client = client_class.from_service_account_info(info) + client = client_class.from_service_account_info(info, transport=transport_name) assert client.transport._credentials == creds assert isinstance(client, client_class) - assert client.transport._host == "bigquerydatatransfer.googleapis.com:443" + assert client.transport._host == ("bigquerydatatransfer.googleapis.com:443") @pytest.mark.parametrize( @@ -140,23 +144,33 @@ def test_data_transfer_service_client_service_account_always_use_jwt( @pytest.mark.parametrize( - "client_class", [DataTransferServiceClient, DataTransferServiceAsyncClient,] + "client_class,transport_name", + [ + (DataTransferServiceClient, "grpc"), + (DataTransferServiceAsyncClient, "grpc_asyncio"), + ], ) -def test_data_transfer_service_client_from_service_account_file(client_class): +def test_data_transfer_service_client_from_service_account_file( + client_class, transport_name +): creds = ga_credentials.AnonymousCredentials() with mock.patch.object( service_account.Credentials, "from_service_account_file" ) as factory: factory.return_value = creds - client = client_class.from_service_account_file("dummy/file/path.json") + client = client_class.from_service_account_file( + "dummy/file/path.json", transport=transport_name + ) assert client.transport._credentials == creds assert isinstance(client, client_class) - client = client_class.from_service_account_json("dummy/file/path.json") + client = client_class.from_service_account_json( + "dummy/file/path.json", transport=transport_name + ) assert client.transport._credentials == creds assert isinstance(client, client_class) - assert client.transport._host == "bigquerydatatransfer.googleapis.com:443" + assert client.transport._host == ("bigquerydatatransfer.googleapis.com:443") def test_data_transfer_service_client_get_transport_class(): @@ -522,7 +536,9 @@ def test_data_transfer_service_client_client_options_scopes( client_class, transport_class, transport_name ): # Check the case scopes are provided. - options = client_options.ClientOptions(scopes=["1", "2"],) + options = client_options.ClientOptions( + scopes=["1", "2"], + ) with mock.patch.object(transport_class, "__init__") as patched: patched.return_value = None client = client_class(client_options=options, transport=transport_name) @@ -662,10 +678,17 @@ def test_data_transfer_service_client_create_channel_credentials_file( ) -@pytest.mark.parametrize("request_type", [datatransfer.GetDataSourceRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + datatransfer.GetDataSourceRequest, + dict, + ], +) def test_get_data_source(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -730,7 +753,8 @@ def test_get_data_source_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -746,7 +770,8 @@ async def test_get_data_source_async( transport: str = "grpc_asyncio", request_type=datatransfer.GetDataSourceRequest ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -823,7 +848,7 @@ def test_get_data_source_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.GetDataSourceRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_data_source), "__call__") as call: @@ -837,7 +862,10 @@ def test_get_data_source_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -850,7 +878,7 @@ async def test_get_data_source_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.GetDataSourceRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_data_source), "__call__") as call: @@ -866,7 +894,10 @@ async def test_get_data_source_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_get_data_source_flattened(): @@ -880,7 +911,9 @@ def test_get_data_source_flattened(): call.return_value = datatransfer.DataSource() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.get_data_source(name="name_value",) + client.get_data_source( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -900,7 +933,8 @@ def test_get_data_source_flattened_error(): # fields is an error. with pytest.raises(ValueError): client.get_data_source( - datatransfer.GetDataSourceRequest(), name="name_value", + datatransfer.GetDataSourceRequest(), + name="name_value", ) @@ -920,7 +954,9 @@ async def test_get_data_source_flattened_async(): ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.get_data_source(name="name_value",) + response = await client.get_data_source( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -941,14 +977,22 @@ async def test_get_data_source_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.get_data_source( - datatransfer.GetDataSourceRequest(), name="name_value", + datatransfer.GetDataSourceRequest(), + name="name_value", ) -@pytest.mark.parametrize("request_type", [datatransfer.ListDataSourcesRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + datatransfer.ListDataSourcesRequest, + dict, + ], +) def test_list_data_sources(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -979,7 +1023,8 @@ def test_list_data_sources_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -997,7 +1042,8 @@ async def test_list_data_sources_async( transport: str = "grpc_asyncio", request_type=datatransfer.ListDataSourcesRequest ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1040,7 +1086,7 @@ def test_list_data_sources_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.ListDataSourcesRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -1056,7 +1102,10 @@ def test_list_data_sources_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -1069,7 +1118,7 @@ async def test_list_data_sources_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.ListDataSourcesRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -1087,7 +1136,10 @@ async def test_list_data_sources_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_data_sources_flattened(): @@ -1103,7 +1155,9 @@ def test_list_data_sources_flattened(): call.return_value = datatransfer.ListDataSourcesResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.list_data_sources(parent="parent_value",) + client.list_data_sources( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -1123,7 +1177,8 @@ def test_list_data_sources_flattened_error(): # fields is an error. with pytest.raises(ValueError): client.list_data_sources( - datatransfer.ListDataSourcesRequest(), parent="parent_value", + datatransfer.ListDataSourcesRequest(), + parent="parent_value", ) @@ -1145,7 +1200,9 @@ async def test_list_data_sources_flattened_async(): ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.list_data_sources(parent="parent_value",) + response = await client.list_data_sources( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -1166,13 +1223,15 @@ async def test_list_data_sources_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.list_data_sources( - datatransfer.ListDataSourcesRequest(), parent="parent_value", + datatransfer.ListDataSourcesRequest(), + parent="parent_value", ) def test_list_data_sources_pager(transport_name: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -1190,13 +1249,20 @@ def test_list_data_sources_pager(transport_name: str = "grpc"): next_page_token="abc", ), datatransfer.ListDataSourcesResponse( - data_sources=[], next_page_token="def", + data_sources=[], + next_page_token="def", ), datatransfer.ListDataSourcesResponse( - data_sources=[datatransfer.DataSource(),], next_page_token="ghi", + data_sources=[ + datatransfer.DataSource(), + ], + next_page_token="ghi", ), datatransfer.ListDataSourcesResponse( - data_sources=[datatransfer.DataSource(), datatransfer.DataSource(),], + data_sources=[ + datatransfer.DataSource(), + datatransfer.DataSource(), + ], ), RuntimeError, ) @@ -1209,14 +1275,15 @@ def test_list_data_sources_pager(transport_name: str = "grpc"): assert pager._metadata == metadata - results = [i for i in pager] + results = list(pager) assert len(results) == 6 assert all(isinstance(i, datatransfer.DataSource) for i in results) def test_list_data_sources_pages(transport_name: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -1234,13 +1301,20 @@ def test_list_data_sources_pages(transport_name: str = "grpc"): next_page_token="abc", ), datatransfer.ListDataSourcesResponse( - data_sources=[], next_page_token="def", + data_sources=[], + next_page_token="def", ), datatransfer.ListDataSourcesResponse( - data_sources=[datatransfer.DataSource(),], next_page_token="ghi", + data_sources=[ + datatransfer.DataSource(), + ], + next_page_token="ghi", ), datatransfer.ListDataSourcesResponse( - data_sources=[datatransfer.DataSource(), datatransfer.DataSource(),], + data_sources=[ + datatransfer.DataSource(), + datatransfer.DataSource(), + ], ), RuntimeError, ) @@ -1272,20 +1346,29 @@ async def test_list_data_sources_async_pager(): next_page_token="abc", ), datatransfer.ListDataSourcesResponse( - data_sources=[], next_page_token="def", + data_sources=[], + next_page_token="def", ), datatransfer.ListDataSourcesResponse( - data_sources=[datatransfer.DataSource(),], next_page_token="ghi", + data_sources=[ + datatransfer.DataSource(), + ], + next_page_token="ghi", ), datatransfer.ListDataSourcesResponse( - data_sources=[datatransfer.DataSource(), datatransfer.DataSource(),], + data_sources=[ + datatransfer.DataSource(), + datatransfer.DataSource(), + ], ), RuntimeError, ) - async_pager = await client.list_data_sources(request={},) + async_pager = await client.list_data_sources( + request={}, + ) assert async_pager.next_page_token == "abc" responses = [] - async for response in async_pager: + async for response in async_pager: # pragma: no branch responses.append(response) assert len(responses) == 6 @@ -1315,29 +1398,43 @@ async def test_list_data_sources_async_pages(): next_page_token="abc", ), datatransfer.ListDataSourcesResponse( - data_sources=[], next_page_token="def", + data_sources=[], + next_page_token="def", ), datatransfer.ListDataSourcesResponse( - data_sources=[datatransfer.DataSource(),], next_page_token="ghi", + data_sources=[ + datatransfer.DataSource(), + ], + next_page_token="ghi", ), datatransfer.ListDataSourcesResponse( - data_sources=[datatransfer.DataSource(), datatransfer.DataSource(),], + data_sources=[ + datatransfer.DataSource(), + datatransfer.DataSource(), + ], ), RuntimeError, ) pages = [] - async for page_ in (await client.list_data_sources(request={})).pages: + async for page_ in ( + await client.list_data_sources(request={}) + ).pages: # pragma: no branch pages.append(page_) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token @pytest.mark.parametrize( - "request_type", [datatransfer.CreateTransferConfigRequest, dict,] + "request_type", + [ + datatransfer.CreateTransferConfigRequest, + dict, + ], ) def test_create_transfer_config(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1387,7 +1484,8 @@ def test_create_transfer_config_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -1406,7 +1504,8 @@ async def test_create_transfer_config_async( request_type=datatransfer.CreateTransferConfigRequest, ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1467,7 +1566,7 @@ def test_create_transfer_config_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.CreateTransferConfigRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -1483,7 +1582,10 @@ def test_create_transfer_config_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -1496,7 +1598,7 @@ async def test_create_transfer_config_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.CreateTransferConfigRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -1514,7 +1616,10 @@ async def test_create_transfer_config_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_create_transfer_config_flattened(): @@ -1614,11 +1719,16 @@ async def test_create_transfer_config_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [datatransfer.UpdateTransferConfigRequest, dict,] + "request_type", + [ + datatransfer.UpdateTransferConfigRequest, + dict, + ], ) def test_update_transfer_config(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1668,7 +1778,8 @@ def test_update_transfer_config_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -1687,7 +1798,8 @@ async def test_update_transfer_config_async( request_type=datatransfer.UpdateTransferConfigRequest, ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1748,7 +1860,7 @@ def test_update_transfer_config_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.UpdateTransferConfigRequest() - request.transfer_config.name = "transfer_config.name/value" + request.transfer_config.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -1766,7 +1878,7 @@ def test_update_transfer_config_field_headers(): _, _, kw = call.mock_calls[0] assert ( "x-goog-request-params", - "transfer_config.name=transfer_config.name/value", + "transfer_config.name=name_value", ) in kw["metadata"] @@ -1780,7 +1892,7 @@ async def test_update_transfer_config_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.UpdateTransferConfigRequest() - request.transfer_config.name = "transfer_config.name/value" + request.transfer_config.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -1800,7 +1912,7 @@ async def test_update_transfer_config_field_headers_async(): _, _, kw = call.mock_calls[0] assert ( "x-goog-request-params", - "transfer_config.name=transfer_config.name/value", + "transfer_config.name=name_value", ) in kw["metadata"] @@ -1901,11 +2013,16 @@ async def test_update_transfer_config_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [datatransfer.DeleteTransferConfigRequest, dict,] + "request_type", + [ + datatransfer.DeleteTransferConfigRequest, + dict, + ], ) def test_delete_transfer_config(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1933,7 +2050,8 @@ def test_delete_transfer_config_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -1952,7 +2070,8 @@ async def test_delete_transfer_config_async( request_type=datatransfer.DeleteTransferConfigRequest, ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -1990,7 +2109,7 @@ def test_delete_transfer_config_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.DeleteTransferConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2006,7 +2125,10 @@ def test_delete_transfer_config_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2019,7 +2141,7 @@ async def test_delete_transfer_config_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.DeleteTransferConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2035,7 +2157,10 @@ async def test_delete_transfer_config_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_delete_transfer_config_flattened(): @@ -2051,7 +2176,9 @@ def test_delete_transfer_config_flattened(): call.return_value = None # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.delete_transfer_config(name="name_value",) + client.delete_transfer_config( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -2071,7 +2198,8 @@ def test_delete_transfer_config_flattened_error(): # fields is an error. with pytest.raises(ValueError): client.delete_transfer_config( - datatransfer.DeleteTransferConfigRequest(), name="name_value", + datatransfer.DeleteTransferConfigRequest(), + name="name_value", ) @@ -2091,7 +2219,9 @@ async def test_delete_transfer_config_flattened_async(): call.return_value = grpc_helpers_async.FakeUnaryUnaryCall(None) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.delete_transfer_config(name="name_value",) + response = await client.delete_transfer_config( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -2112,14 +2242,22 @@ async def test_delete_transfer_config_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.delete_transfer_config( - datatransfer.DeleteTransferConfigRequest(), name="name_value", + datatransfer.DeleteTransferConfigRequest(), + name="name_value", ) -@pytest.mark.parametrize("request_type", [datatransfer.GetTransferConfigRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + datatransfer.GetTransferConfigRequest, + dict, + ], +) def test_get_transfer_config(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2169,7 +2307,8 @@ def test_get_transfer_config_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2187,7 +2326,8 @@ async def test_get_transfer_config_async( transport: str = "grpc_asyncio", request_type=datatransfer.GetTransferConfigRequest ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2248,7 +2388,7 @@ def test_get_transfer_config_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.GetTransferConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2264,7 +2404,10 @@ def test_get_transfer_config_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2277,7 +2420,7 @@ async def test_get_transfer_config_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.GetTransferConfigRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2295,7 +2438,10 @@ async def test_get_transfer_config_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_get_transfer_config_flattened(): @@ -2311,7 +2457,9 @@ def test_get_transfer_config_flattened(): call.return_value = transfer.TransferConfig() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.get_transfer_config(name="name_value",) + client.get_transfer_config( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -2331,7 +2479,8 @@ def test_get_transfer_config_flattened_error(): # fields is an error. with pytest.raises(ValueError): client.get_transfer_config( - datatransfer.GetTransferConfigRequest(), name="name_value", + datatransfer.GetTransferConfigRequest(), + name="name_value", ) @@ -2353,7 +2502,9 @@ async def test_get_transfer_config_flattened_async(): ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.get_transfer_config(name="name_value",) + response = await client.get_transfer_config( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -2374,16 +2525,22 @@ async def test_get_transfer_config_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.get_transfer_config( - datatransfer.GetTransferConfigRequest(), name="name_value", + datatransfer.GetTransferConfigRequest(), + name="name_value", ) @pytest.mark.parametrize( - "request_type", [datatransfer.ListTransferConfigsRequest, dict,] + "request_type", + [ + datatransfer.ListTransferConfigsRequest, + dict, + ], ) def test_list_transfer_configs(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2414,7 +2571,8 @@ def test_list_transfer_configs_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2433,7 +2591,8 @@ async def test_list_transfer_configs_async( request_type=datatransfer.ListTransferConfigsRequest, ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2476,7 +2635,7 @@ def test_list_transfer_configs_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.ListTransferConfigsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2492,7 +2651,10 @@ def test_list_transfer_configs_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2505,7 +2667,7 @@ async def test_list_transfer_configs_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.ListTransferConfigsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2523,7 +2685,10 @@ async def test_list_transfer_configs_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_transfer_configs_flattened(): @@ -2539,7 +2704,9 @@ def test_list_transfer_configs_flattened(): call.return_value = datatransfer.ListTransferConfigsResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.list_transfer_configs(parent="parent_value",) + client.list_transfer_configs( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -2559,7 +2726,8 @@ def test_list_transfer_configs_flattened_error(): # fields is an error. with pytest.raises(ValueError): client.list_transfer_configs( - datatransfer.ListTransferConfigsRequest(), parent="parent_value", + datatransfer.ListTransferConfigsRequest(), + parent="parent_value", ) @@ -2581,7 +2749,9 @@ async def test_list_transfer_configs_flattened_async(): ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.list_transfer_configs(parent="parent_value",) + response = await client.list_transfer_configs( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -2602,13 +2772,15 @@ async def test_list_transfer_configs_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.list_transfer_configs( - datatransfer.ListTransferConfigsRequest(), parent="parent_value", + datatransfer.ListTransferConfigsRequest(), + parent="parent_value", ) def test_list_transfer_configs_pager(transport_name: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2626,10 +2798,14 @@ def test_list_transfer_configs_pager(transport_name: str = "grpc"): next_page_token="abc", ), datatransfer.ListTransferConfigsResponse( - transfer_configs=[], next_page_token="def", + transfer_configs=[], + next_page_token="def", ), datatransfer.ListTransferConfigsResponse( - transfer_configs=[transfer.TransferConfig(),], next_page_token="ghi", + transfer_configs=[ + transfer.TransferConfig(), + ], + next_page_token="ghi", ), datatransfer.ListTransferConfigsResponse( transfer_configs=[ @@ -2648,14 +2824,15 @@ def test_list_transfer_configs_pager(transport_name: str = "grpc"): assert pager._metadata == metadata - results = [i for i in pager] + results = list(pager) assert len(results) == 6 assert all(isinstance(i, transfer.TransferConfig) for i in results) def test_list_transfer_configs_pages(transport_name: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2673,10 +2850,14 @@ def test_list_transfer_configs_pages(transport_name: str = "grpc"): next_page_token="abc", ), datatransfer.ListTransferConfigsResponse( - transfer_configs=[], next_page_token="def", + transfer_configs=[], + next_page_token="def", ), datatransfer.ListTransferConfigsResponse( - transfer_configs=[transfer.TransferConfig(),], next_page_token="ghi", + transfer_configs=[ + transfer.TransferConfig(), + ], + next_page_token="ghi", ), datatransfer.ListTransferConfigsResponse( transfer_configs=[ @@ -2714,10 +2895,14 @@ async def test_list_transfer_configs_async_pager(): next_page_token="abc", ), datatransfer.ListTransferConfigsResponse( - transfer_configs=[], next_page_token="def", + transfer_configs=[], + next_page_token="def", ), datatransfer.ListTransferConfigsResponse( - transfer_configs=[transfer.TransferConfig(),], next_page_token="ghi", + transfer_configs=[ + transfer.TransferConfig(), + ], + next_page_token="ghi", ), datatransfer.ListTransferConfigsResponse( transfer_configs=[ @@ -2727,10 +2912,12 @@ async def test_list_transfer_configs_async_pager(): ), RuntimeError, ) - async_pager = await client.list_transfer_configs(request={},) + async_pager = await client.list_transfer_configs( + request={}, + ) assert async_pager.next_page_token == "abc" responses = [] - async for response in async_pager: + async for response in async_pager: # pragma: no branch responses.append(response) assert len(responses) == 6 @@ -2760,10 +2947,14 @@ async def test_list_transfer_configs_async_pages(): next_page_token="abc", ), datatransfer.ListTransferConfigsResponse( - transfer_configs=[], next_page_token="def", + transfer_configs=[], + next_page_token="def", ), datatransfer.ListTransferConfigsResponse( - transfer_configs=[transfer.TransferConfig(),], next_page_token="ghi", + transfer_configs=[ + transfer.TransferConfig(), + ], + next_page_token="ghi", ), datatransfer.ListTransferConfigsResponse( transfer_configs=[ @@ -2774,18 +2965,25 @@ async def test_list_transfer_configs_async_pages(): RuntimeError, ) pages = [] - async for page_ in (await client.list_transfer_configs(request={})).pages: + async for page_ in ( + await client.list_transfer_configs(request={}) + ).pages: # pragma: no branch pages.append(page_) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token @pytest.mark.parametrize( - "request_type", [datatransfer.ScheduleTransferRunsRequest, dict,] + "request_type", + [ + datatransfer.ScheduleTransferRunsRequest, + dict, + ], ) def test_schedule_transfer_runs(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2813,7 +3011,8 @@ def test_schedule_transfer_runs_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -2832,7 +3031,8 @@ async def test_schedule_transfer_runs_async( request_type=datatransfer.ScheduleTransferRunsRequest, ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -2872,7 +3072,7 @@ def test_schedule_transfer_runs_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.ScheduleTransferRunsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2888,7 +3088,10 @@ def test_schedule_transfer_runs_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -2901,7 +3104,7 @@ async def test_schedule_transfer_runs_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.ScheduleTransferRunsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -2919,7 +3122,10 @@ async def test_schedule_transfer_runs_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_schedule_transfer_runs_flattened(): @@ -3029,11 +3235,16 @@ async def test_schedule_transfer_runs_flattened_error_async(): @pytest.mark.parametrize( - "request_type", [datatransfer.StartManualTransferRunsRequest, dict,] + "request_type", + [ + datatransfer.StartManualTransferRunsRequest, + dict, + ], ) def test_start_manual_transfer_runs(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3061,7 +3272,8 @@ def test_start_manual_transfer_runs_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3080,7 +3292,8 @@ async def test_start_manual_transfer_runs_async( request_type=datatransfer.StartManualTransferRunsRequest, ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3120,7 +3333,7 @@ def test_start_manual_transfer_runs_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.StartManualTransferRunsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -3136,7 +3349,10 @@ def test_start_manual_transfer_runs_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -3149,7 +3365,7 @@ async def test_start_manual_transfer_runs_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.StartManualTransferRunsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -3167,13 +3383,23 @@ async def test_start_manual_transfer_runs_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] -@pytest.mark.parametrize("request_type", [datatransfer.GetTransferRunRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + datatransfer.GetTransferRunRequest, + dict, + ], +) def test_get_transfer_run(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3213,7 +3439,8 @@ def test_get_transfer_run_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3229,7 +3456,8 @@ async def test_get_transfer_run_async( transport: str = "grpc_asyncio", request_type=datatransfer.GetTransferRunRequest ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3280,7 +3508,7 @@ def test_get_transfer_run_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.GetTransferRunRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_transfer_run), "__call__") as call: @@ -3294,7 +3522,10 @@ def test_get_transfer_run_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -3307,7 +3538,7 @@ async def test_get_transfer_run_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.GetTransferRunRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_transfer_run), "__call__") as call: @@ -3323,7 +3554,10 @@ async def test_get_transfer_run_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_get_transfer_run_flattened(): @@ -3337,7 +3571,9 @@ def test_get_transfer_run_flattened(): call.return_value = transfer.TransferRun() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.get_transfer_run(name="name_value",) + client.get_transfer_run( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -3357,7 +3593,8 @@ def test_get_transfer_run_flattened_error(): # fields is an error. with pytest.raises(ValueError): client.get_transfer_run( - datatransfer.GetTransferRunRequest(), name="name_value", + datatransfer.GetTransferRunRequest(), + name="name_value", ) @@ -3377,7 +3614,9 @@ async def test_get_transfer_run_flattened_async(): ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.get_transfer_run(name="name_value",) + response = await client.get_transfer_run( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -3398,14 +3637,22 @@ async def test_get_transfer_run_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.get_transfer_run( - datatransfer.GetTransferRunRequest(), name="name_value", + datatransfer.GetTransferRunRequest(), + name="name_value", ) -@pytest.mark.parametrize("request_type", [datatransfer.DeleteTransferRunRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + datatransfer.DeleteTransferRunRequest, + dict, + ], +) def test_delete_transfer_run(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3433,7 +3680,8 @@ def test_delete_transfer_run_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3451,7 +3699,8 @@ async def test_delete_transfer_run_async( transport: str = "grpc_asyncio", request_type=datatransfer.DeleteTransferRunRequest ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3489,7 +3738,7 @@ def test_delete_transfer_run_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.DeleteTransferRunRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -3505,7 +3754,10 @@ def test_delete_transfer_run_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -3518,7 +3770,7 @@ async def test_delete_transfer_run_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.DeleteTransferRunRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -3534,7 +3786,10 @@ async def test_delete_transfer_run_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_delete_transfer_run_flattened(): @@ -3550,7 +3805,9 @@ def test_delete_transfer_run_flattened(): call.return_value = None # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.delete_transfer_run(name="name_value",) + client.delete_transfer_run( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -3570,7 +3827,8 @@ def test_delete_transfer_run_flattened_error(): # fields is an error. with pytest.raises(ValueError): client.delete_transfer_run( - datatransfer.DeleteTransferRunRequest(), name="name_value", + datatransfer.DeleteTransferRunRequest(), + name="name_value", ) @@ -3590,7 +3848,9 @@ async def test_delete_transfer_run_flattened_async(): call.return_value = grpc_helpers_async.FakeUnaryUnaryCall(None) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.delete_transfer_run(name="name_value",) + response = await client.delete_transfer_run( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -3611,14 +3871,22 @@ async def test_delete_transfer_run_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.delete_transfer_run( - datatransfer.DeleteTransferRunRequest(), name="name_value", + datatransfer.DeleteTransferRunRequest(), + name="name_value", ) -@pytest.mark.parametrize("request_type", [datatransfer.ListTransferRunsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + datatransfer.ListTransferRunsRequest, + dict, + ], +) def test_list_transfer_runs(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3649,7 +3917,8 @@ def test_list_transfer_runs_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3667,7 +3936,8 @@ async def test_list_transfer_runs_async( transport: str = "grpc_asyncio", request_type=datatransfer.ListTransferRunsRequest ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -3710,7 +3980,7 @@ def test_list_transfer_runs_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.ListTransferRunsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -3726,7 +3996,10 @@ def test_list_transfer_runs_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -3739,7 +4012,7 @@ async def test_list_transfer_runs_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.ListTransferRunsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -3757,7 +4030,10 @@ async def test_list_transfer_runs_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_transfer_runs_flattened(): @@ -3773,7 +4049,9 @@ def test_list_transfer_runs_flattened(): call.return_value = datatransfer.ListTransferRunsResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.list_transfer_runs(parent="parent_value",) + client.list_transfer_runs( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -3793,7 +4071,8 @@ def test_list_transfer_runs_flattened_error(): # fields is an error. with pytest.raises(ValueError): client.list_transfer_runs( - datatransfer.ListTransferRunsRequest(), parent="parent_value", + datatransfer.ListTransferRunsRequest(), + parent="parent_value", ) @@ -3815,7 +4094,9 @@ async def test_list_transfer_runs_flattened_async(): ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.list_transfer_runs(parent="parent_value",) + response = await client.list_transfer_runs( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -3836,13 +4117,15 @@ async def test_list_transfer_runs_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.list_transfer_runs( - datatransfer.ListTransferRunsRequest(), parent="parent_value", + datatransfer.ListTransferRunsRequest(), + parent="parent_value", ) def test_list_transfer_runs_pager(transport_name: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3860,13 +4143,20 @@ def test_list_transfer_runs_pager(transport_name: str = "grpc"): next_page_token="abc", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[], next_page_token="def", + transfer_runs=[], + next_page_token="def", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[transfer.TransferRun(),], next_page_token="ghi", + transfer_runs=[ + transfer.TransferRun(), + ], + next_page_token="ghi", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[transfer.TransferRun(), transfer.TransferRun(),], + transfer_runs=[ + transfer.TransferRun(), + transfer.TransferRun(), + ], ), RuntimeError, ) @@ -3879,14 +4169,15 @@ def test_list_transfer_runs_pager(transport_name: str = "grpc"): assert pager._metadata == metadata - results = [i for i in pager] + results = list(pager) assert len(results) == 6 assert all(isinstance(i, transfer.TransferRun) for i in results) def test_list_transfer_runs_pages(transport_name: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -3904,13 +4195,20 @@ def test_list_transfer_runs_pages(transport_name: str = "grpc"): next_page_token="abc", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[], next_page_token="def", + transfer_runs=[], + next_page_token="def", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[transfer.TransferRun(),], next_page_token="ghi", + transfer_runs=[ + transfer.TransferRun(), + ], + next_page_token="ghi", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[transfer.TransferRun(), transfer.TransferRun(),], + transfer_runs=[ + transfer.TransferRun(), + transfer.TransferRun(), + ], ), RuntimeError, ) @@ -3942,20 +4240,29 @@ async def test_list_transfer_runs_async_pager(): next_page_token="abc", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[], next_page_token="def", + transfer_runs=[], + next_page_token="def", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[transfer.TransferRun(),], next_page_token="ghi", + transfer_runs=[ + transfer.TransferRun(), + ], + next_page_token="ghi", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[transfer.TransferRun(), transfer.TransferRun(),], + transfer_runs=[ + transfer.TransferRun(), + transfer.TransferRun(), + ], ), RuntimeError, ) - async_pager = await client.list_transfer_runs(request={},) + async_pager = await client.list_transfer_runs( + request={}, + ) assert async_pager.next_page_token == "abc" responses = [] - async for response in async_pager: + async for response in async_pager: # pragma: no branch responses.append(response) assert len(responses) == 6 @@ -3985,27 +4292,43 @@ async def test_list_transfer_runs_async_pages(): next_page_token="abc", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[], next_page_token="def", + transfer_runs=[], + next_page_token="def", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[transfer.TransferRun(),], next_page_token="ghi", + transfer_runs=[ + transfer.TransferRun(), + ], + next_page_token="ghi", ), datatransfer.ListTransferRunsResponse( - transfer_runs=[transfer.TransferRun(), transfer.TransferRun(),], + transfer_runs=[ + transfer.TransferRun(), + transfer.TransferRun(), + ], ), RuntimeError, ) pages = [] - async for page_ in (await client.list_transfer_runs(request={})).pages: + async for page_ in ( + await client.list_transfer_runs(request={}) + ).pages: # pragma: no branch pages.append(page_) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token -@pytest.mark.parametrize("request_type", [datatransfer.ListTransferLogsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + datatransfer.ListTransferLogsRequest, + dict, + ], +) def test_list_transfer_logs(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4036,7 +4359,8 @@ def test_list_transfer_logs_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4054,7 +4378,8 @@ async def test_list_transfer_logs_async( transport: str = "grpc_asyncio", request_type=datatransfer.ListTransferLogsRequest ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4097,7 +4422,7 @@ def test_list_transfer_logs_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.ListTransferLogsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -4113,7 +4438,10 @@ def test_list_transfer_logs_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4126,7 +4454,7 @@ async def test_list_transfer_logs_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.ListTransferLogsRequest() - request.parent = "parent/value" + request.parent = "parent_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -4144,7 +4472,10 @@ async def test_list_transfer_logs_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "parent=parent_value", + ) in kw["metadata"] def test_list_transfer_logs_flattened(): @@ -4160,7 +4491,9 @@ def test_list_transfer_logs_flattened(): call.return_value = datatransfer.ListTransferLogsResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.list_transfer_logs(parent="parent_value",) + client.list_transfer_logs( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -4180,7 +4513,8 @@ def test_list_transfer_logs_flattened_error(): # fields is an error. with pytest.raises(ValueError): client.list_transfer_logs( - datatransfer.ListTransferLogsRequest(), parent="parent_value", + datatransfer.ListTransferLogsRequest(), + parent="parent_value", ) @@ -4202,7 +4536,9 @@ async def test_list_transfer_logs_flattened_async(): ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.list_transfer_logs(parent="parent_value",) + response = await client.list_transfer_logs( + parent="parent_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -4223,13 +4559,15 @@ async def test_list_transfer_logs_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.list_transfer_logs( - datatransfer.ListTransferLogsRequest(), parent="parent_value", + datatransfer.ListTransferLogsRequest(), + parent="parent_value", ) def test_list_transfer_logs_pager(transport_name: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4247,10 +4585,14 @@ def test_list_transfer_logs_pager(transport_name: str = "grpc"): next_page_token="abc", ), datatransfer.ListTransferLogsResponse( - transfer_messages=[], next_page_token="def", + transfer_messages=[], + next_page_token="def", ), datatransfer.ListTransferLogsResponse( - transfer_messages=[transfer.TransferMessage(),], next_page_token="ghi", + transfer_messages=[ + transfer.TransferMessage(), + ], + next_page_token="ghi", ), datatransfer.ListTransferLogsResponse( transfer_messages=[ @@ -4269,14 +4611,15 @@ def test_list_transfer_logs_pager(transport_name: str = "grpc"): assert pager._metadata == metadata - results = [i for i in pager] + results = list(pager) assert len(results) == 6 assert all(isinstance(i, transfer.TransferMessage) for i in results) def test_list_transfer_logs_pages(transport_name: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials, transport=transport_name, + credentials=ga_credentials.AnonymousCredentials, + transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4294,10 +4637,14 @@ def test_list_transfer_logs_pages(transport_name: str = "grpc"): next_page_token="abc", ), datatransfer.ListTransferLogsResponse( - transfer_messages=[], next_page_token="def", + transfer_messages=[], + next_page_token="def", ), datatransfer.ListTransferLogsResponse( - transfer_messages=[transfer.TransferMessage(),], next_page_token="ghi", + transfer_messages=[ + transfer.TransferMessage(), + ], + next_page_token="ghi", ), datatransfer.ListTransferLogsResponse( transfer_messages=[ @@ -4335,10 +4682,14 @@ async def test_list_transfer_logs_async_pager(): next_page_token="abc", ), datatransfer.ListTransferLogsResponse( - transfer_messages=[], next_page_token="def", + transfer_messages=[], + next_page_token="def", ), datatransfer.ListTransferLogsResponse( - transfer_messages=[transfer.TransferMessage(),], next_page_token="ghi", + transfer_messages=[ + transfer.TransferMessage(), + ], + next_page_token="ghi", ), datatransfer.ListTransferLogsResponse( transfer_messages=[ @@ -4348,10 +4699,12 @@ async def test_list_transfer_logs_async_pager(): ), RuntimeError, ) - async_pager = await client.list_transfer_logs(request={},) + async_pager = await client.list_transfer_logs( + request={}, + ) assert async_pager.next_page_token == "abc" responses = [] - async for response in async_pager: + async for response in async_pager: # pragma: no branch responses.append(response) assert len(responses) == 6 @@ -4381,10 +4734,14 @@ async def test_list_transfer_logs_async_pages(): next_page_token="abc", ), datatransfer.ListTransferLogsResponse( - transfer_messages=[], next_page_token="def", + transfer_messages=[], + next_page_token="def", ), datatransfer.ListTransferLogsResponse( - transfer_messages=[transfer.TransferMessage(),], next_page_token="ghi", + transfer_messages=[ + transfer.TransferMessage(), + ], + next_page_token="ghi", ), datatransfer.ListTransferLogsResponse( transfer_messages=[ @@ -4395,16 +4752,25 @@ async def test_list_transfer_logs_async_pages(): RuntimeError, ) pages = [] - async for page_ in (await client.list_transfer_logs(request={})).pages: + async for page_ in ( + await client.list_transfer_logs(request={}) + ).pages: # pragma: no branch pages.append(page_) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token -@pytest.mark.parametrize("request_type", [datatransfer.CheckValidCredsRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + datatransfer.CheckValidCredsRequest, + dict, + ], +) def test_check_valid_creds(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4416,7 +4782,9 @@ def test_check_valid_creds(request_type, transport: str = "grpc"): type(client.transport.check_valid_creds), "__call__" ) as call: # Designate an appropriate return value for the call. - call.return_value = datatransfer.CheckValidCredsResponse(has_valid_creds=True,) + call.return_value = datatransfer.CheckValidCredsResponse( + has_valid_creds=True, + ) response = client.check_valid_creds(request) # Establish that the underlying gRPC stub method was called. @@ -4433,7 +4801,8 @@ def test_check_valid_creds_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4451,7 +4820,8 @@ async def test_check_valid_creds_async( transport: str = "grpc_asyncio", request_type=datatransfer.CheckValidCredsRequest ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4464,7 +4834,9 @@ async def test_check_valid_creds_async( ) as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( - datatransfer.CheckValidCredsResponse(has_valid_creds=True,) + datatransfer.CheckValidCredsResponse( + has_valid_creds=True, + ) ) response = await client.check_valid_creds(request) @@ -4492,7 +4864,7 @@ def test_check_valid_creds_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.CheckValidCredsRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -4508,7 +4880,10 @@ def test_check_valid_creds_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4521,7 +4896,7 @@ async def test_check_valid_creds_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.CheckValidCredsRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -4539,7 +4914,10 @@ async def test_check_valid_creds_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_check_valid_creds_flattened(): @@ -4555,7 +4933,9 @@ def test_check_valid_creds_flattened(): call.return_value = datatransfer.CheckValidCredsResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - client.check_valid_creds(name="name_value",) + client.check_valid_creds( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -4575,7 +4955,8 @@ def test_check_valid_creds_flattened_error(): # fields is an error. with pytest.raises(ValueError): client.check_valid_creds( - datatransfer.CheckValidCredsRequest(), name="name_value", + datatransfer.CheckValidCredsRequest(), + name="name_value", ) @@ -4597,7 +4978,9 @@ async def test_check_valid_creds_flattened_async(): ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. - response = await client.check_valid_creds(name="name_value",) + response = await client.check_valid_creds( + name="name_value", + ) # Establish that the underlying call was made with the expected # request object values. @@ -4618,14 +5001,22 @@ async def test_check_valid_creds_flattened_error_async(): # fields is an error. with pytest.raises(ValueError): await client.check_valid_creds( - datatransfer.CheckValidCredsRequest(), name="name_value", + datatransfer.CheckValidCredsRequest(), + name="name_value", ) -@pytest.mark.parametrize("request_type", [datatransfer.EnrollDataSourcesRequest, dict,]) +@pytest.mark.parametrize( + "request_type", + [ + datatransfer.EnrollDataSourcesRequest, + dict, + ], +) def test_enroll_data_sources(request_type, transport: str = "grpc"): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4653,7 +5044,8 @@ def test_enroll_data_sources_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. @@ -4671,7 +5063,8 @@ async def test_enroll_data_sources_async( transport: str = "grpc_asyncio", request_type=datatransfer.EnrollDataSourcesRequest ): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, @@ -4709,7 +5102,7 @@ def test_enroll_data_sources_field_headers(): # a field header. Set these to a non-empty value. request = datatransfer.EnrollDataSourcesRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -4725,7 +5118,10 @@ def test_enroll_data_sources_field_headers(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] @pytest.mark.asyncio @@ -4738,7 +5134,7 @@ async def test_enroll_data_sources_field_headers_async(): # a field header. Set these to a non-empty value. request = datatransfer.EnrollDataSourcesRequest() - request.name = "name/value" + request.name = "name_value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( @@ -4754,7 +5150,10 @@ async def test_enroll_data_sources_field_headers_async(): # Establish that the field header was sent. _, _, kw = call.mock_calls[0] - assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] + assert ( + "x-goog-request-params", + "name=name_value", + ) in kw["metadata"] def test_credentials_transport_error(): @@ -4764,7 +5163,8 @@ def test_credentials_transport_error(): ) with pytest.raises(ValueError): client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), transport=transport, + credentials=ga_credentials.AnonymousCredentials(), + transport=transport, ) # It is an error to provide a credentials file and a transport instance. @@ -4784,7 +5184,10 @@ def test_credentials_transport_error(): options = client_options.ClientOptions() options.api_key = "api_key" with pytest.raises(ValueError): - client = DataTransferServiceClient(client_options=options, transport=transport,) + client = DataTransferServiceClient( + client_options=options, + transport=transport, + ) # It is an error to provide an api_key and a credential. options = mock.Mock() @@ -4800,7 +5203,8 @@ def test_credentials_transport_error(): ) with pytest.raises(ValueError): client = DataTransferServiceClient( - client_options={"scopes": ["1", "2"]}, transport=transport, + client_options={"scopes": ["1", "2"]}, + transport=transport, ) @@ -4843,12 +5247,28 @@ def test_transport_adc(transport_class): adc.assert_called_once() +@pytest.mark.parametrize( + "transport_name", + [ + "grpc", + ], +) +def test_transport_kind(transport_name): + transport = DataTransferServiceClient.get_transport_class(transport_name)( + credentials=ga_credentials.AnonymousCredentials(), + ) + assert transport.kind == transport_name + + def test_transport_grpc_default(): # A client should use the gRPC transport by default. client = DataTransferServiceClient( credentials=ga_credentials.AnonymousCredentials(), ) - assert isinstance(client.transport, transports.DataTransferServiceGrpcTransport,) + assert isinstance( + client.transport, + transports.DataTransferServiceGrpcTransport, + ) def test_data_transfer_service_base_transport_error(): @@ -4896,6 +5316,14 @@ def test_data_transfer_service_base_transport(): with pytest.raises(NotImplementedError): transport.close() + # Catch all for all remaining methods and properties + remainder = [ + "kind", + ] + for r in remainder: + with pytest.raises(NotImplementedError): + getattr(transport, r)() + def test_data_transfer_service_base_transport_with_credentials_file(): # Instantiate the base transport with a credentials file @@ -4907,7 +5335,8 @@ def test_data_transfer_service_base_transport_with_credentials_file(): Transport.return_value = None load_creds.return_value = (ga_credentials.AnonymousCredentials(), None) transport = transports.DataTransferServiceTransport( - credentials_file="credentials.json", quota_project_id="octopus", + credentials_file="credentials.json", + quota_project_id="octopus", ) load_creds.assert_called_once_with( "credentials.json", @@ -5042,24 +5471,40 @@ def test_data_transfer_service_grpc_transport_client_cert_source_for_mtls( ) -def test_data_transfer_service_host_no_port(): +@pytest.mark.parametrize( + "transport_name", + [ + "grpc", + "grpc_asyncio", + ], +) +def test_data_transfer_service_host_no_port(transport_name): client = DataTransferServiceClient( credentials=ga_credentials.AnonymousCredentials(), client_options=client_options.ClientOptions( api_endpoint="bigquerydatatransfer.googleapis.com" ), + transport=transport_name, ) - assert client.transport._host == "bigquerydatatransfer.googleapis.com:443" + assert client.transport._host == ("bigquerydatatransfer.googleapis.com:443") -def test_data_transfer_service_host_with_port(): +@pytest.mark.parametrize( + "transport_name", + [ + "grpc", + "grpc_asyncio", + ], +) +def test_data_transfer_service_host_with_port(transport_name): client = DataTransferServiceClient( credentials=ga_credentials.AnonymousCredentials(), client_options=client_options.ClientOptions( api_endpoint="bigquerydatatransfer.googleapis.com:8000" ), + transport=transport_name, ) - assert client.transport._host == "bigquerydatatransfer.googleapis.com:8000" + assert client.transport._host == ("bigquerydatatransfer.googleapis.com:8000") def test_data_transfer_service_grpc_transport_channel(): @@ -5067,7 +5512,8 @@ def test_data_transfer_service_grpc_transport_channel(): # Check that channel is used if provided. transport = transports.DataTransferServiceGrpcTransport( - host="squid.clam.whelk", channel=channel, + host="squid.clam.whelk", + channel=channel, ) assert transport.grpc_channel == channel assert transport._host == "squid.clam.whelk:443" @@ -5079,7 +5525,8 @@ def test_data_transfer_service_grpc_asyncio_transport_channel(): # Check that channel is used if provided. transport = transports.DataTransferServiceGrpcAsyncIOTransport( - host="squid.clam.whelk", channel=channel, + host="squid.clam.whelk", + channel=channel, ) assert transport.grpc_channel == channel assert transport._host == "squid.clam.whelk:443" @@ -5190,7 +5637,8 @@ def test_data_source_path(): project = "squid" data_source = "clam" expected = "projects/{project}/dataSources/{data_source}".format( - project=project, data_source=data_source, + project=project, + data_source=data_source, ) actual = DataTransferServiceClient.data_source_path(project, data_source) assert expected == actual @@ -5213,7 +5661,9 @@ def test_run_path(): transfer_config = "nudibranch" run = "cuttlefish" expected = "projects/{project}/transferConfigs/{transfer_config}/runs/{run}".format( - project=project, transfer_config=transfer_config, run=run, + project=project, + transfer_config=transfer_config, + run=run, ) actual = DataTransferServiceClient.run_path(project, transfer_config, run) assert expected == actual @@ -5236,7 +5686,8 @@ def test_transfer_config_path(): project = "scallop" transfer_config = "abalone" expected = "projects/{project}/transferConfigs/{transfer_config}".format( - project=project, transfer_config=transfer_config, + project=project, + transfer_config=transfer_config, ) actual = DataTransferServiceClient.transfer_config_path(project, transfer_config) assert expected == actual @@ -5276,7 +5727,9 @@ def test_parse_common_billing_account_path(): def test_common_folder_path(): folder = "oyster" - expected = "folders/{folder}".format(folder=folder,) + expected = "folders/{folder}".format( + folder=folder, + ) actual = DataTransferServiceClient.common_folder_path(folder) assert expected == actual @@ -5294,7 +5747,9 @@ def test_parse_common_folder_path(): def test_common_organization_path(): organization = "cuttlefish" - expected = "organizations/{organization}".format(organization=organization,) + expected = "organizations/{organization}".format( + organization=organization, + ) actual = DataTransferServiceClient.common_organization_path(organization) assert expected == actual @@ -5312,7 +5767,9 @@ def test_parse_common_organization_path(): def test_common_project_path(): project = "winkle" - expected = "projects/{project}".format(project=project,) + expected = "projects/{project}".format( + project=project, + ) actual = DataTransferServiceClient.common_project_path(project) assert expected == actual @@ -5332,7 +5789,8 @@ def test_common_location_path(): project = "scallop" location = "abalone" expected = "projects/{project}/locations/{location}".format( - project=project, location=location, + project=project, + location=location, ) actual = DataTransferServiceClient.common_location_path(project, location) assert expected == actual @@ -5357,7 +5815,8 @@ def test_client_with_default_client_info(): transports.DataTransferServiceTransport, "_prep_wrapped_messages" ) as prep: client = DataTransferServiceClient( - credentials=ga_credentials.AnonymousCredentials(), client_info=client_info, + credentials=ga_credentials.AnonymousCredentials(), + client_info=client_info, ) prep.assert_called_once_with(client_info) @@ -5366,7 +5825,8 @@ def test_client_with_default_client_info(): ) as prep: transport_class = DataTransferServiceClient.get_transport_class() transport = transport_class( - credentials=ga_credentials.AnonymousCredentials(), client_info=client_info, + credentials=ga_credentials.AnonymousCredentials(), + client_info=client_info, ) prep.assert_called_once_with(client_info) @@ -5374,7 +5834,8 @@ def test_client_with_default_client_info(): @pytest.mark.asyncio async def test_transport_close_async(): client = DataTransferServiceAsyncClient( - credentials=ga_credentials.AnonymousCredentials(), transport="grpc_asyncio", + credentials=ga_credentials.AnonymousCredentials(), + transport="grpc_asyncio", ) with mock.patch.object( type(getattr(client.transport, "grpc_channel")), "close" diff --git a/tests/unit/test_shim.py b/tests/unit/test_shim.py index bf80d60b..c5c20db2 100644 --- a/tests/unit/test_shim.py +++ b/tests/unit/test_shim.py @@ -18,8 +18,7 @@ def test_shim(): - from google.cloud import bigquery_datatransfer - from google.cloud import bigquery_datatransfer_v1 + from google.cloud import bigquery_datatransfer, bigquery_datatransfer_v1 assert sorted(bigquery_datatransfer.__all__) == sorted( bigquery_datatransfer_v1.__all__