@@ -530,6 +530,10 @@ def test_deletion_of_cache_dir_faulty_download(self, patch):
530530 datasets_cache_dir = os .path .join (openml .config .get_cache_directory (), "datasets" )
531531 assert len (os .listdir (datasets_cache_dir )) == 0
532532
533+ @pytest .mark .skipif (
534+ os .getenv ("OPENML_USE_LOCAL_SERVICES" ) == "true" ,
535+ reason = "Pending resolution of #1657" ,
536+ )
533537 @pytest .mark .test_server ()
534538 def test_publish_dataset (self ):
535539 arff_file_path = self .static_cache_dir / "org" / "openml" / "test" / "datasets" / "2" / "dataset.arff"
@@ -566,6 +570,10 @@ def test__retrieve_class_labels(self):
566570 labels = custom_ds .retrieve_class_labels (target_name = custom_ds .features [31 ].name )
567571 assert labels == ["COIL" , "SHEET" ]
568572
573+ @pytest .mark .skipif (
574+ os .getenv ("OPENML_USE_LOCAL_SERVICES" ) == "true" ,
575+ reason = "Pending resolution of #1657" ,
576+ )
569577 @pytest .mark .test_server ()
570578 def test_upload_dataset_with_url (self ):
571579 dataset = OpenMLDataset (
@@ -689,6 +697,10 @@ def test_attributes_arff_from_df_unknown_dtype(self):
689697 with pytest .raises (ValueError , match = err_msg ):
690698 attributes_arff_from_df (df )
691699
700+ @pytest .mark .skipif (
701+ os .getenv ("OPENML_USE_LOCAL_SERVICES" ) == "true" ,
702+ reason = "Pending resolution of #1657" ,
703+ )
692704 @pytest .mark .test_server ()
693705 def test_create_dataset_numpy (self ):
694706 data = np .array ([[1 , 2 , 3 ], [1.2 , 2.5 , 3.8 ], [2 , 5 , 8 ], [0 , 1 , 0 ]]).T
@@ -723,6 +735,10 @@ def test_create_dataset_numpy(self):
723735 ), "Uploaded arff does not match original one"
724736 assert _get_online_dataset_format (dataset .id ) == "arff" , "Wrong format for dataset"
725737
738+ @pytest .mark .skipif (
739+ os .getenv ("OPENML_USE_LOCAL_SERVICES" ) == "true" ,
740+ reason = "Pending resolution of #1657" ,
741+ )
726742 @pytest .mark .test_server ()
727743 def test_create_dataset_list (self ):
728744 data = [
@@ -778,6 +794,10 @@ def test_create_dataset_list(self):
778794 ), "Uploaded ARFF does not match original one"
779795 assert _get_online_dataset_format (dataset .id ) == "arff" , "Wrong format for dataset"
780796
797+ @pytest .mark .skipif (
798+ os .getenv ("OPENML_USE_LOCAL_SERVICES" ) == "true" ,
799+ reason = "Pending resolution of #1657" ,
800+ )
781801 @pytest .mark .test_server ()
782802 def test_create_dataset_sparse (self ):
783803 # test the scipy.sparse.coo_matrix
@@ -926,6 +946,10 @@ def test_get_online_dataset_format(self):
926946 dataset_id
927947 ), "The format of the ARFF files is different"
928948
949+ @pytest .mark .skipif (
950+ os .getenv ("OPENML_USE_LOCAL_SERVICES" ) == "true" ,
951+ reason = "Pending resolution of #1657" ,
952+ )
929953 @pytest .mark .test_server ()
930954 def test_create_dataset_pandas (self ):
931955 data = [
@@ -1151,6 +1175,10 @@ def test_ignore_attributes_dataset(self):
11511175 paper_url = paper_url ,
11521176 )
11531177
1178+ @pytest .mark .skipif (
1179+ os .getenv ("OPENML_USE_LOCAL_SERVICES" ) == "true" ,
1180+ reason = "Pending resolution of #1657" ,
1181+ )
11541182 @pytest .mark .test_server ()
11551183 def test_publish_fetch_ignore_attribute (self ):
11561184 """Test to upload and retrieve dataset and check ignore_attributes"""
@@ -1270,6 +1298,10 @@ def test_create_dataset_row_id_attribute_error(self):
12701298 paper_url = paper_url ,
12711299 )
12721300
1301+ @pytest .mark .skipif (
1302+ os .getenv ("OPENML_USE_LOCAL_SERVICES" ) == "true" ,
1303+ reason = "Pending resolution of #1657" ,
1304+ )
12731305 @pytest .mark .test_server ()
12741306 def test_create_dataset_row_id_attribute_inference (self ):
12751307 # meta-information
@@ -1438,6 +1470,10 @@ def test_data_edit_non_critical_field(self):
14381470 edited_dataset = openml .datasets .get_dataset (did )
14391471 assert edited_dataset .description == desc
14401472
1473+ @pytest .mark .skipif (
1474+ os .getenv ("OPENML_USE_LOCAL_SERVICES" ) == "true" ,
1475+ reason = "Pending resolution of #1657" ,
1476+ )
14411477 @pytest .mark .test_server ()
14421478 def test_data_edit_critical_field (self ):
14431479 # Case 2
@@ -1490,6 +1526,10 @@ def test_data_edit_requires_valid_dataset(self):
14901526 description = "xor operation dataset" ,
14911527 )
14921528
1529+ @pytest .mark .skipif (
1530+ os .getenv ("OPENML_USE_LOCAL_SERVICES" ) == "true" ,
1531+ reason = "Pending resolution of #1657" ,
1532+ )
14931533 @pytest .mark .test_server ()
14941534 def test_data_edit_cannot_edit_critical_field_if_dataset_has_task (self ):
14951535 # Need to own a dataset to be able to edit meta-data
@@ -2008,4 +2048,4 @@ def test_get_dataset_parquet(requests_mock, test_files_directory):
20082048 assert dataset ._parquet_url is not None
20092049 assert dataset .parquet_file is not None
20102050 assert os .path .isfile (dataset .parquet_file )
2011- assert dataset .data_file is None # is alias for arff path
2051+ assert dataset .data_file is None # is alias for arff path
0 commit comments