@@ -247,12 +247,14 @@ def test_create_dataset(self):
247
247
def test_get_dataset (self ):
248
248
dataset_id = _make_dataset_id ("get_dataset" )
249
249
client = Config .CLIENT
250
- dataset_arg = Dataset (client .dataset (dataset_id ))
250
+ project = client .project
251
+ dataset_ref = bigquery .DatasetReference (project , dataset_id )
252
+ dataset_arg = Dataset (dataset_ref )
251
253
dataset_arg .friendly_name = "Friendly"
252
254
dataset_arg .description = "Description"
253
255
dataset = retry_403 (client .create_dataset )(dataset_arg )
254
256
self .to_delete .append (dataset )
255
- dataset_ref = client . dataset ( dataset_id )
257
+ dataset_ref = bigquery . DatasetReference ( project , dataset_id )
256
258
257
259
# Get with a reference.
258
260
got = client .get_dataset (dataset_ref )
@@ -416,17 +418,18 @@ def test_create_table_w_time_partitioning_w_clustering_fields(self):
416
418
417
419
def test_delete_dataset_with_string (self ):
418
420
dataset_id = _make_dataset_id ("delete_table_true" )
419
- dataset_ref = Config .CLIENT .dataset (dataset_id )
421
+ project = Config .CLIENT .project
422
+ dataset_ref = bigquery .DatasetReference (project , dataset_id )
420
423
retry_403 (Config .CLIENT .create_dataset )(Dataset (dataset_ref ))
421
424
self .assertTrue (_dataset_exists (dataset_ref ))
422
425
Config .CLIENT .delete_dataset (dataset_id )
423
426
self .assertFalse (_dataset_exists (dataset_ref ))
424
427
425
428
def test_delete_dataset_delete_contents_true (self ):
426
429
dataset_id = _make_dataset_id ("delete_table_true" )
427
- dataset = retry_403 ( Config .CLIENT .create_dataset )(
428
- Dataset ( Config . CLIENT . dataset ( dataset_id ) )
429
- )
430
+ project = Config .CLIENT .project
431
+ dataset_ref = bigquery . DatasetReference ( project , dataset_id )
432
+ dataset = retry_403 ( Config . CLIENT . create_dataset )( Dataset ( dataset_ref ) )
430
433
431
434
table_id = "test_table"
432
435
table_arg = Table (dataset .table (table_id ), schema = SCHEMA )
@@ -1363,7 +1366,9 @@ def test_extract_table(self):
1363
1366
source_blob_name = "person_ages.csv"
1364
1367
dataset_id = _make_dataset_id ("load_gcs_then_extract" )
1365
1368
table_id = "test_table"
1366
- table_ref = Config .CLIENT .dataset (dataset_id ).table (table_id )
1369
+ project = Config .CLIENT .project
1370
+ dataset_ref = bigquery .DatasetReference (project , dataset_id )
1371
+ table_ref = dataset_ref .table (table_id )
1367
1372
table = Table (table_ref )
1368
1373
self .to_delete .insert (0 , table )
1369
1374
bucket = self ._create_bucket (bucket_name )
@@ -1546,8 +1551,10 @@ def test_query_w_wrong_config(self):
1546
1551
rows = list (Config .CLIENT .query ("SELECT 1;" ).result ())
1547
1552
assert rows [0 ][0 ] == 1
1548
1553
1554
+ project = Config .CLIENT .project
1555
+ dataset_ref = bigquery .DatasetReference (project , "dset" )
1549
1556
bad_config = LoadJobConfig ()
1550
- bad_config .destination = Config . CLIENT . dataset ( "dset" ) .table ("tbl" )
1557
+ bad_config .destination = dataset_ref .table ("tbl" )
1551
1558
with self .assertRaises (Exception ):
1552
1559
Config .CLIENT .query (good_query , job_config = bad_config ).result ()
1553
1560
@@ -2678,7 +2685,9 @@ def test_list_rows_max_results_w_bqstorage(self):
2678
2685
self .assertEqual (len (dataframe .index ), 100 )
2679
2686
2680
2687
def temp_dataset (self , dataset_id , location = None ):
2681
- dataset = Dataset (Config .CLIENT .dataset (dataset_id ))
2688
+ project = Config .CLIENT .project
2689
+ dataset_ref = bigquery .DatasetReference (project , dataset_id )
2690
+ dataset = Dataset (dataset_ref )
2682
2691
if location :
2683
2692
dataset .location = location
2684
2693
dataset = retry_403 (Config .CLIENT .create_dataset )(dataset )
0 commit comments