Skip to content

Commit

Permalink
bigquery: remove client from Dataset (#4018)
Browse files Browse the repository at this point in the history
  • Loading branch information
jba authored and tswast committed Oct 12, 2017
1 parent 97f4e1c commit a8169a4
Show file tree
Hide file tree
Showing 7 changed files with 49 additions and 87 deletions.
8 changes: 4 additions & 4 deletions bigquery/google/cloud/bigquery/client.py
Original file line number Diff line number Diff line change
Expand Up @@ -190,7 +190,7 @@ def create_dataset(self, dataset):
path = '/projects/%s/datasets' % (dataset.project,)
api_response = self._connection.api_request(
method='POST', path=path, data=dataset._build_resource())
return Dataset.from_api_repr(api_response, self)
return Dataset.from_api_repr(api_response)

def get_dataset(self, dataset_ref):
"""Fetch the dataset referenced by ``dataset_ref``
Expand All @@ -204,7 +204,7 @@ def get_dataset(self, dataset_ref):
"""
api_response = self._connection.api_request(
method='GET', path=dataset_ref.path)
return Dataset.from_api_repr(api_response, self)
return Dataset.from_api_repr(api_response)

def get_table(self, table_ref):
"""Fetch the table referenced by ``table_ref``
Expand Down Expand Up @@ -262,7 +262,7 @@ def update_dataset(self, dataset, fields):
headers = None
api_response = self._connection.api_request(
method='PATCH', path=path, data=partial, headers=headers)
return Dataset.from_api_repr(api_response, self)
return Dataset.from_api_repr(api_response)

def list_dataset_tables(self, dataset, max_results=None, page_token=None):
"""List tables in the dataset.
Expand Down Expand Up @@ -622,7 +622,7 @@ def _item_to_dataset(iterator, resource):
:rtype: :class:`.Dataset`
:returns: The next dataset in the page.
"""
return Dataset.from_api_repr(resource, iterator.client)
return Dataset.from_api_repr(resource)


def _item_to_job(iterator, resource):
Expand Down
27 changes: 8 additions & 19 deletions bigquery/google/cloud/bigquery/dataset.py
Original file line number Diff line number Diff line change
Expand Up @@ -157,38 +157,31 @@ class Dataset(object):
:type dataset_id: str
:param dataset_id: the ID of the dataset
:type client: :class:`google.cloud.bigquery.client.Client`
:param client: (Optional) A client which holds credentials and project
configuration for the dataset (which requires a project).
:type access_entries: list of :class:`AccessEntry`
:param access_entries: roles granted to entities for this dataset
:type project: str
:param project: (Optional) project ID for the dataset (defaults to
the project of the client).
:param project: (Optional) project ID for the dataset.
"""

_access_entries = None

def __init__(self,
dataset_id,
client=None,
access_entries=(),
project=None):
self._dataset_id = dataset_id
self._client = client
self._properties = {}
# Let the @property do validation.
self.access_entries = access_entries
self._project = project or (client and client.project)
self._project = project

@property
def project(self):
"""Project bound to the dataset.
:rtype: str
:returns: the project (derived from the client).
:returns: the project.
"""
return self._project

Expand Down Expand Up @@ -373,25 +366,21 @@ def location(self, value):
self._properties['location'] = value

@classmethod
def from_api_repr(cls, resource, client):
def from_api_repr(cls, resource):
"""Factory: construct a dataset given its API representation
:type resource: dict
:param resource: dataset resource representation returned from the API
:type client: :class:`google.cloud.bigquery.client.Client`
:param client: Client which holds credentials and project
configuration for the dataset.
:rtype: :class:`google.cloud.bigquery.dataset.Dataset`
:returns: Dataset parsed from ``resource``.
"""
if ('datasetReference' not in resource or
'datasetId' not in resource['datasetReference']):
dsr = resource.get('datasetReference')
if dsr is None or 'datasetId' not in dsr:
raise KeyError('Resource lacks required identity information:'
'["datasetReference"]["datasetId"]')
dataset_id = resource['datasetReference']['datasetId']
dataset = cls(dataset_id, client=client)
dataset_id = dsr['datasetId']
dataset = cls(dataset_id, project=dsr['projectId'])
dataset._set_properties(resource)
return dataset

Expand Down
19 changes: 10 additions & 9 deletions bigquery/google/cloud/bigquery/job.py
Original file line number Diff line number Diff line change
Expand Up @@ -842,7 +842,8 @@ def from_api_repr(cls, resource, client):
"""
job_id, config = cls._get_resource_config(resource)
dest_config = config['destinationTable']
dataset = Dataset(dest_config['datasetId'], client)
dataset = Dataset(dest_config['datasetId'],
project=dest_config['projectId'])
table_ref = TableReference(dataset, dest_config['tableId'])
destination = Table(table_ref, client=client)
source_urls = config.get('sourceUris', ())
Expand Down Expand Up @@ -958,7 +959,8 @@ def from_api_repr(cls, resource, client):
"""
job_id, config = cls._get_resource_config(resource)
dest_config = config['destinationTable']
dataset = Dataset(dest_config['datasetId'], client)
dataset = Dataset(dest_config['datasetId'],
project=dest_config['projectId'])
table_ref = TableReference(dataset, dest_config['tableId'])
destination = Table(table_ref, client=client)
sources = []
Expand All @@ -970,7 +972,8 @@ def from_api_repr(cls, resource, client):
"Resource missing 'sourceTables' / 'sourceTable'")
source_configs = [single]
for source_config in source_configs:
dataset = Dataset(source_config['datasetId'], client)
dataset = Dataset(source_config['datasetId'],
project=source_config['projectId'])
table_ref = TableReference(dataset, source_config['tableId'])
sources.append(Table(table_ref, client=client))
job = cls(job_id, destination, sources, client=client)
Expand Down Expand Up @@ -1423,17 +1426,16 @@ def _copy_configuration_properties(self, configuration):
dest_local = self._destination_table_resource()
if dest_remote != dest_local:
project = dest_remote['projectId']
dataset = Dataset(
dest_remote['datasetId'], self._client, project=project)
dataset = Dataset(dest_remote['datasetId'], project=project)
self.destination = dataset.table(dest_remote['tableId'])

def_ds = configuration.get('defaultDataset')
if def_ds is None:
if self.default_dataset is not None:
del self.default_dataset
else:
project = def_ds['projectId']
self.default_dataset = Dataset(def_ds['datasetId'], self._client)
self.default_dataset = Dataset(def_ds['datasetId'],
project=def_ds['projectId'])

udf_resources = []
for udf_mapping in configuration.get(self._UDF_KEY, ()):
Expand Down Expand Up @@ -1579,7 +1581,6 @@ def referenced_tables(self):
if the query has not yet completed.
"""
tables = []
client = self._require_client(None)
datasets_by_project_name = {}

for table in self._job_statistics().get('referencedTables', ()):
Expand All @@ -1589,7 +1590,7 @@ def referenced_tables(self):
ds_name = table['datasetId']
t_dataset = datasets_by_project_name.get((t_project, ds_name))
if t_dataset is None:
t_dataset = Dataset(ds_name, client, project=t_project)
t_dataset = Dataset(ds_name, project=t_project)
datasets_by_project_name[(t_project, ds_name)] = t_dataset

t_name = table['tableId']
Expand Down
5 changes: 3 additions & 2 deletions bigquery/tests/system.py
Original file line number Diff line number Diff line change
Expand Up @@ -117,11 +117,12 @@ def test_create_dataset(self):

self.assertTrue(_dataset_exists(dataset))
self.assertEqual(dataset.dataset_id, DATASET_ID)
self.assertEqual(dataset.project, Config.CLIENT.project)

def test_get_dataset(self):
DATASET_ID = _make_dataset_id('get_dataset')
client = Config.CLIENT
dataset_arg = Dataset(DATASET_ID)
dataset_arg = Dataset(DATASET_ID, project=client.project)
dataset_arg.friendly_name = 'Friendly'
dataset_arg.description = 'Description'
dataset = retry_403(client.create_dataset)(dataset_arg)
Expand Down Expand Up @@ -1195,7 +1196,7 @@ def test_dump_table_w_public_data(self):
DATASET_ID = 'samples'
TABLE_NAME = 'natality'

dataset = Dataset(DATASET_ID, Config.CLIENT, project=PUBLIC)
dataset = Dataset(DATASET_ID, project=PUBLIC)
table_ref = dataset.table(TABLE_NAME)
table = Config.CLIENT.get_table(table_ref)
self._fetch_single_page(table)
Expand Down
69 changes: 20 additions & 49 deletions bigquery/tests/unit/test_dataset.py
Original file line number Diff line number Diff line change
Expand Up @@ -210,11 +210,9 @@ def _verify_resource_properties(self, dataset, resource):
self.assertEqual(dataset.access_entries, [])

def test_ctor_defaults(self):
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID, project=self.PROJECT)
self.assertEqual(dataset.dataset_id, self.DS_ID)
self.assertIs(dataset._client, client)
self.assertEqual(dataset.project, client.project)
self.assertEqual(dataset.project, self.PROJECT)
self.assertEqual(
dataset.path,
'/projects/%s/datasets/%s' % (self.PROJECT, self.DS_ID))
Expand All @@ -238,12 +236,10 @@ def test_ctor_explicit(self):
bharney = AccessEntry('OWNER', 'userByEmail', '[email protected]')
entries = [phred, bharney]
OTHER_PROJECT = 'foo-bar-123'
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client,
dataset = self._make_one(self.DS_ID,
access_entries=entries,
project=OTHER_PROJECT)
self.assertEqual(dataset.dataset_id, self.DS_ID)
self.assertIs(dataset._client, client)
self.assertEqual(dataset.project, OTHER_PROJECT)
self.assertEqual(
dataset.path,
Expand All @@ -262,89 +258,76 @@ def test_ctor_explicit(self):
self.assertIsNone(dataset.location)

def test_access_entries_setter_non_list(self):
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID)
with self.assertRaises(TypeError):
dataset.access_entries = object()

def test_access_entries_setter_invalid_field(self):
from google.cloud.bigquery.dataset import AccessEntry

client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID)
phred = AccessEntry('OWNER', 'userByEmail', '[email protected]')
with self.assertRaises(ValueError):
dataset.access_entries = [phred, object()]

def test_access_entries_setter(self):
from google.cloud.bigquery.dataset import AccessEntry

client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID)
phred = AccessEntry('OWNER', 'userByEmail', '[email protected]')
bharney = AccessEntry('OWNER', 'userByEmail', '[email protected]')
dataset.access_entries = [phred, bharney]
self.assertEqual(dataset.access_entries, [phred, bharney])

def test_default_table_expiration_ms_setter_bad_value(self):
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID)
with self.assertRaises(ValueError):
dataset.default_table_expiration_ms = 'bogus'

def test_default_table_expiration_ms_setter(self):
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID)
dataset.default_table_expiration_ms = 12345
self.assertEqual(dataset.default_table_expiration_ms, 12345)

def test_description_setter_bad_value(self):
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID)
with self.assertRaises(ValueError):
dataset.description = 12345

def test_description_setter(self):
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID)
dataset.description = 'DESCRIPTION'
self.assertEqual(dataset.description, 'DESCRIPTION')

def test_friendly_name_setter_bad_value(self):
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID)
with self.assertRaises(ValueError):
dataset.friendly_name = 12345

def test_friendly_name_setter(self):
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID)
dataset.friendly_name = 'FRIENDLY'
self.assertEqual(dataset.friendly_name, 'FRIENDLY')

def test_location_setter_bad_value(self):
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID)
with self.assertRaises(ValueError):
dataset.location = 12345

def test_location_setter(self):
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client)
dataset = self._make_one(self.DS_ID)
dataset.location = 'LOCATION'
self.assertEqual(dataset.location, 'LOCATION')

def test_from_api_repr_missing_identity(self):
self._setUpConstants()
client = _Client(self.PROJECT)
RESOURCE = {}
klass = self._get_target_class()
with self.assertRaises(KeyError):
klass.from_api_repr(RESOURCE, client=client)
klass.from_api_repr(RESOURCE)

def test_from_api_repr_bare(self):
self._setUpConstants()
client = _Client(self.PROJECT)
RESOURCE = {
'id': '%s:%s' % (self.PROJECT, self.DS_ID),
'datasetReference': {
Expand All @@ -353,24 +336,20 @@ def test_from_api_repr_bare(self):
}
}
klass = self._get_target_class()
dataset = klass.from_api_repr(RESOURCE, client=client)
self.assertIs(dataset._client, client)
dataset = klass.from_api_repr(RESOURCE)
self._verify_resource_properties(dataset, RESOURCE)

def test_from_api_repr_w_properties(self):
client = _Client(self.PROJECT)
RESOURCE = self._makeResource()
klass = self._get_target_class()
dataset = klass.from_api_repr(RESOURCE, client=client)
self.assertIs(dataset._client, client)
dataset = klass.from_api_repr(RESOURCE)
self._verify_resource_properties(dataset, RESOURCE)

def test__parse_access_entries_w_unknown_entity_type(self):
ACCESS = [
{'role': 'READER', 'unknown': 'UNKNOWN'},
]
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client=client)
dataset = self._make_one(self.DS_ID)
with self.assertRaises(ValueError):
dataset._parse_access_entries(ACCESS)

Expand All @@ -383,24 +362,16 @@ def test__parse_access_entries_w_extra_keys(self):
'userByEmail': USER_EMAIL,
},
]
client = _Client(self.PROJECT)
dataset = self._make_one(self.DS_ID, client=client)
dataset = self._make_one(self.DS_ID)
with self.assertRaises(ValueError):
dataset._parse_access_entries(ACCESS)

def test_table(self):
from google.cloud.bigquery.table import TableReference

client = _Client(project=self.PROJECT)
dataset = self._make_one(self.DS_ID, client=client)
dataset = self._make_one(self.DS_ID, project=self.PROJECT)
table = dataset.table('table_id')
self.assertIsInstance(table, TableReference)
self.assertEqual(table.table_id, 'table_id')
self.assertEqual(table.dataset_id, self.DS_ID)
self.assertEqual(table.project, self.PROJECT)


class _Client(object):

def __init__(self, project='project'):
self.project = project
Loading

0 comments on commit a8169a4

Please sign in to comment.