|
19 | 19 | from google.cloud.bigquery._http import Connection |
20 | 20 | from google.cloud.bigquery.dataset import Dataset |
21 | 21 | from google.cloud.bigquery.job import CopyJob |
22 | | -from google.cloud.bigquery.job import ExtractTableToStorageJob |
23 | | -from google.cloud.bigquery.job import LoadTableFromStorageJob |
| 22 | +from google.cloud.bigquery.job import ExtractJob |
| 23 | +from google.cloud.bigquery.job import LoadJob |
24 | 24 | from google.cloud.bigquery.job import QueryJob |
25 | 25 | from google.cloud.bigquery.query import QueryResults |
26 | 26 |
|
@@ -169,20 +169,20 @@ def job_from_resource(self, resource): |
169 | 169 | :param resource: one job resource from API response |
170 | 170 |
|
171 | 171 | :rtype: One of: |
172 | | - :class:`google.cloud.bigquery.job.LoadTableFromStorageJob`, |
| 172 | + :class:`google.cloud.bigquery.job.LoadJob`, |
173 | 173 | :class:`google.cloud.bigquery.job.CopyJob`, |
174 | | - :class:`google.cloud.bigquery.job.ExtractTableToStorageJob`, |
| 174 | + :class:`google.cloud.bigquery.job.ExtractJob`, |
175 | 175 | :class:`google.cloud.bigquery.job.QueryJob`, |
176 | 176 | :class:`google.cloud.bigquery.job.RunSyncQueryJob` |
177 | 177 | :returns: the job instance, constructed via the resource |
178 | 178 | """ |
179 | 179 | config = resource['configuration'] |
180 | 180 | if 'load' in config: |
181 | | - return LoadTableFromStorageJob.from_api_repr(resource, self) |
| 181 | + return LoadJob.from_api_repr(resource, self) |
182 | 182 | elif 'copy' in config: |
183 | 183 | return CopyJob.from_api_repr(resource, self) |
184 | 184 | elif 'extract' in config: |
185 | | - return ExtractTableToStorageJob.from_api_repr(resource, self) |
| 185 | + return ExtractJob.from_api_repr(resource, self) |
186 | 186 | elif 'query' in config: |
187 | 187 | return QueryJob.from_api_repr(resource, self) |
188 | 188 | raise ValueError('Cannot parse job resource') |
@@ -253,11 +253,10 @@ def load_table_from_storage(self, job_name, destination, *source_uris): |
253 | 253 | :param source_uris: URIs of data files to be loaded; in format |
254 | 254 | ``gs://<bucket_name>/<object_name_or_glob>``. |
255 | 255 |
|
256 | | - :rtype: :class:`google.cloud.bigquery.job.LoadTableFromStorageJob` |
257 | | - :returns: a new ``LoadTableFromStorageJob`` instance |
| 256 | + :rtype: :class:`google.cloud.bigquery.job.LoadJob` |
| 257 | + :returns: a new ``LoadJob`` instance |
258 | 258 | """ |
259 | | - return LoadTableFromStorageJob(job_name, destination, source_uris, |
260 | | - client=self) |
| 259 | + return LoadJob(job_name, destination, source_uris, client=self) |
261 | 260 |
|
262 | 261 | def copy_table(self, job_name, destination, *sources): |
263 | 262 | """Construct a job for copying one or more tables into another table. |
@@ -296,11 +295,10 @@ def extract_table_to_storage(self, job_name, source, *destination_uris): |
296 | 295 | table data is to be extracted; in format |
297 | 296 | ``gs://<bucket_name>/<object_name_or_glob>``. |
298 | 297 |
|
299 | | - :rtype: :class:`google.cloud.bigquery.job.ExtractTableToStorageJob` |
300 | | - :returns: a new ``ExtractTableToStorageJob`` instance |
| 298 | + :rtype: :class:`google.cloud.bigquery.job.ExtractJob` |
| 299 | + :returns: a new ``ExtractJob`` instance |
301 | 300 | """ |
302 | | - return ExtractTableToStorageJob(job_name, source, destination_uris, |
303 | | - client=self) |
| 301 | + return ExtractJob(job_name, source, destination_uris, client=self) |
304 | 302 |
|
305 | 303 | def run_async_query(self, job_name, query, |
306 | 304 | udf_resources=(), query_parameters=()): |
|
0 commit comments