Skip to content

Commit 6f7f171

Browse files
authored
rename datasets-server to dataset-viewer (#6785)
* rename datasets-server to dataset-viewer * use renamed error name
1 parent c3ddb1e commit 6f7f171

File tree

4 files changed

+20
-20
lines changed

4 files changed

+20
-20
lines changed

src/datasets/config.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -193,7 +193,7 @@
193193
)
194194
TIME_OUT_REMOTE_CODE = 15
195195

196-
# Datasets-server
196+
# Dataset viewer API
197197
USE_PARQUET_EXPORT = True
198198

199199
# Batch size constants. For more info, see:

src/datasets/load.py

+6-6
Original file line numberDiff line numberDiff line change
@@ -71,7 +71,7 @@
7171
_hash_python_lines,
7272
)
7373
from .splits import Split
74-
from .utils import _datasets_server
74+
from .utils import _dataset_viewer
7575
from .utils.deprecation_utils import deprecated
7676
from .utils.file_utils import (
7777
OfflineModeIsEnabled,
@@ -1236,7 +1236,7 @@ def get_module(self) -> DatasetModule:
12361236
metadata_configs = MetadataConfigs.from_dataset_card_data(dataset_card_data)
12371237
dataset_infos = DatasetInfosDict.from_dataset_card_data(dataset_card_data)
12381238
try:
1239-
exported_dataset_infos = _datasets_server.get_exported_dataset_infos(
1239+
exported_dataset_infos = _dataset_viewer.get_exported_dataset_infos(
12401240
dataset=self.name, revision=self.revision, token=self.download_config.token
12411241
)
12421242
exported_dataset_infos = DatasetInfosDict(
@@ -1245,7 +1245,7 @@ def get_module(self) -> DatasetModule:
12451245
for config_name in exported_dataset_infos
12461246
}
12471247
)
1248-
except _datasets_server.DatasetsServerError:
1248+
except _dataset_viewer.DatasetViewerError:
12491249
exported_dataset_infos = None
12501250
if exported_dataset_infos:
12511251
exported_dataset_infos.update(dataset_infos)
@@ -1372,10 +1372,10 @@ def __init__(
13721372
increase_load_count(name, resource_type="dataset")
13731373

13741374
def get_module(self) -> DatasetModule:
1375-
exported_parquet_files = _datasets_server.get_exported_parquet_files(
1375+
exported_parquet_files = _dataset_viewer.get_exported_parquet_files(
13761376
dataset=self.name, revision=self.revision, token=self.download_config.token
13771377
)
1378-
exported_dataset_infos = _datasets_server.get_exported_dataset_infos(
1378+
exported_dataset_infos = _dataset_viewer.get_exported_dataset_infos(
13791379
dataset=self.name, revision=self.revision, token=self.download_config.token
13801380
)
13811381
dataset_infos = DatasetInfosDict(
@@ -1864,7 +1864,7 @@ def dataset_module_factory(
18641864
return HubDatasetModuleFactoryWithParquetExport(
18651865
path, download_config=download_config, revision=dataset_info.sha
18661866
).get_module()
1867-
except _datasets_server.DatasetsServerError:
1867+
except _dataset_viewer.DatasetViewerError:
18681868
pass
18691869
# Otherwise we must use the dataset script if the user trusts it
18701870
return HubDatasetModuleFactoryWithScript(

src/datasets/utils/_datasets_server.py renamed to src/datasets/utils/_dataset_viewer.py

+11-11
Original file line numberDiff line numberDiff line change
@@ -12,10 +12,10 @@
1212
logger = get_logger(__name__)
1313

1414

15-
class DatasetsServerError(DatasetsError):
16-
"""Dataset-server error.
15+
class DatasetViewerError(DatasetsError):
16+
"""Dataset viewer error.
1717
18-
Raised when trying to use the Datasets-server HTTP API and when trying to access:
18+
Raised when trying to use the dataset viewer HTTP API and when trying to access:
1919
- a missing dataset, or
2020
- a private/gated dataset and the user is not authenticated.
2121
- unavailable /parquet or /info responses
@@ -27,10 +27,10 @@ def get_exported_parquet_files(dataset: str, revision: str, token: Optional[Unio
2727
Get the dataset exported parquet files
2828
Docs: https://huggingface.co/docs/datasets-server/parquet
2929
"""
30-
datasets_server_parquet_url = config.HF_ENDPOINT.replace("://", "://datasets-server.") + "/parquet?dataset="
30+
dataset_viewer_parquet_url = config.HF_ENDPOINT.replace("://", "://datasets-server.") + "/parquet?dataset="
3131
try:
3232
parquet_data_files_response = http_get(
33-
url=datasets_server_parquet_url + dataset,
33+
url=dataset_viewer_parquet_url + dataset,
3434
temp_file=None,
3535
headers=get_authentication_headers_for_url(config.HF_ENDPOINT + f"datasets/{dataset}", token=token),
3636
timeout=100.0,
@@ -53,9 +53,9 @@ def get_exported_parquet_files(dataset: str, revision: str, token: Optional[Unio
5353
logger.debug(
5454
f"Parquet export for {dataset} is available but outdated (revision='{parquet_data_files_response.headers['X-Revision']}')"
5555
)
56-
except Exception as e: # noqa catch any exception of the datasets-server and consider the parquet export doesn't exist
56+
except Exception as e: # noqa catch any exception of the dataset viewer API and consider the parquet export doesn't exist
5757
logger.debug(f"No parquet export for {dataset} available ({type(e).__name__}: {e})")
58-
raise DatasetsServerError("No exported Parquet files available.")
58+
raise DatasetViewerError("No exported Parquet files available.")
5959

6060

6161
def get_exported_dataset_infos(
@@ -65,10 +65,10 @@ def get_exported_dataset_infos(
6565
Get the dataset information, can be useful to get e.g. the dataset features.
6666
Docs: https://huggingface.co/docs/datasets-server/info
6767
"""
68-
datasets_server_info_url = config.HF_ENDPOINT.replace("://", "://datasets-server.") + "/info?dataset="
68+
dataset_viewer_info_url = config.HF_ENDPOINT.replace("://", "://datasets-server.") + "/info?dataset="
6969
try:
7070
info_response = http_get(
71-
url=datasets_server_info_url + dataset,
71+
url=dataset_viewer_info_url + dataset,
7272
temp_file=None,
7373
headers=get_authentication_headers_for_url(config.HF_ENDPOINT + f"datasets/{dataset}", token=token),
7474
timeout=100.0,
@@ -91,6 +91,6 @@ def get_exported_dataset_infos(
9191
logger.debug(
9292
f"Dataset info for {dataset} is available but outdated (revision='{info_response.headers['X-Revision']}')"
9393
)
94-
except Exception as e: # noqa catch any exception of the datasets-server and consider the dataset info doesn't exist
94+
except Exception as e: # noqa catch any exception of the dataset viewer API and consider the dataset info doesn't exist
9595
logger.debug(f"No dataset info for {dataset} available ({type(e).__name__}: {e})")
96-
raise DatasetsServerError("No exported dataset infos available.")
96+
raise DatasetViewerError("No exported dataset infos available.")

tests/test_load.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@
4646
from datasets.packaged_modules.audiofolder.audiofolder import AudioFolder, AudioFolderConfig
4747
from datasets.packaged_modules.imagefolder.imagefolder import ImageFolder, ImageFolderConfig
4848
from datasets.packaged_modules.parquet.parquet import ParquetConfig
49-
from datasets.utils import _datasets_server
49+
from datasets.utils import _dataset_viewer
5050
from datasets.utils.logging import INFO, get_logger
5151

5252
from .utils import (
@@ -862,7 +862,7 @@ def test_HubDatasetModuleFactoryWithParquetExport_errors_on_wrong_sha(self):
862862
download_config=self.download_config,
863863
revision="wrong_sha",
864864
)
865-
with self.assertRaises(_datasets_server.DatasetsServerError):
865+
with self.assertRaises(_dataset_viewer.DatasetViewerError):
866866
factory.get_module()
867867

868868
@pytest.mark.integration

0 commit comments

Comments
 (0)