|
import os |
|
import datasets |
|
from datasets import Features, Value |
|
from huggingface_hub import snapshot_download |
|
import glob |
|
import yaml |
|
|
|
|
|
class PathoBenchConfig(datasets.BuilderConfig): |
|
def __init__(self, **kwargs): |
|
|
|
|
|
self.task_in_dataset = kwargs.pop("task_in_dataset", None) |
|
self.dataset_to_download = kwargs.pop("dataset_to_download", None) |
|
self.force_download = kwargs.pop("force_download", True) |
|
|
|
|
|
if self.dataset_to_download is None and self.task_in_dataset is None: |
|
|
|
self.dataset_to_download = '*' |
|
self.task_in_dataset = '*' |
|
elif self.dataset_to_download is None and self.task_in_dataset is not None: |
|
|
|
raise AssertionError("Dataset needs to be defined for the task_in_dataset provided.") |
|
elif self.dataset_to_download is not None and self.task_in_dataset is None: |
|
|
|
self.task_in_dataset = '*' |
|
|
|
super().__init__(**kwargs) |
|
|
|
|
|
class PathoBenchDataset(datasets.GeneratorBasedBuilder): |
|
""" |
|
Downloads only the .tsv and .yaml files needed to construct the dataset. |
|
Excludes .png images so they don't break the builder. |
|
""" |
|
BUILDER_CONFIGS = [ |
|
PathoBenchConfig(name="custom_config", version="1.0.0", description="PathoBench config") |
|
] |
|
BUILDER_CONFIG_CLASS = PathoBenchConfig |
|
|
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
description="PathoBench: collection of canonical computational pathology tasks", |
|
homepage="https://github.com/mahmoodlab/patho-bench", |
|
license="CC BY-NC-SA 4.0 Deed", |
|
features=Features({ |
|
'path': Value('string') |
|
}) |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
repo_id = "MahmoodLab/patho-bench" |
|
dataset_to_download = self.config.dataset_to_download |
|
local_dir = self._cache_dir_root |
|
force_download = self.config.force_download |
|
task_in_dataset = self.config.task_in_dataset |
|
|
|
|
|
os.makedirs(local_dir, exist_ok=True) |
|
|
|
|
|
snapshot_download( |
|
repo_id=repo_id, |
|
allow_patterns=["available_splits.yaml"], |
|
repo_type="dataset", |
|
local_dir=local_dir, |
|
force_download=force_download, |
|
) |
|
|
|
|
|
with open(os.path.join(local_dir, "available_splits.yaml"), 'r') as file: |
|
available_splits = yaml.safe_load(file) |
|
|
|
|
|
if dataset_to_download != "*": |
|
assert dataset_to_download in available_splits, ( |
|
f"{dataset_to_download} was not found. " |
|
f"Available splits: {list(available_splits.keys())}" |
|
) |
|
if task_in_dataset != "*": |
|
assert task_in_dataset in available_splits[dataset_to_download], ( |
|
f"{task_in_dataset} was not found in {dataset_to_download}. " |
|
f"Available tasks: {available_splits[dataset_to_download]}" |
|
) |
|
|
|
|
|
|
|
|
|
|
|
if dataset_to_download == "*": |
|
|
|
allow_patterns = [ |
|
"**/*.tsv", |
|
"**/about.yaml", |
|
"**/config.yaml", |
|
"available_splits.yaml" |
|
] |
|
else: |
|
if task_in_dataset == "*": |
|
allow_patterns = [ |
|
f"{dataset_to_download}/**/*.tsv", |
|
f"{dataset_to_download}/**/about.yaml", |
|
f"{dataset_to_download}/**/config.yaml", |
|
"available_splits.yaml" |
|
] |
|
else: |
|
allow_patterns = [ |
|
f"{dataset_to_download}/{task_in_dataset}/*.tsv", |
|
f"{dataset_to_download}/{task_in_dataset}/config.yaml", |
|
f"{dataset_to_download}/about.yaml", |
|
"available_splits.yaml" |
|
] |
|
|
|
|
|
snapshot_download( |
|
repo_id=repo_id, |
|
allow_patterns=allow_patterns, |
|
repo_type="dataset", |
|
local_dir=local_dir, |
|
force_download=force_download, |
|
) |
|
|
|
|
|
search_pattern = os.path.join(local_dir, '**', '*.tsv') |
|
all_tsv_splits = glob.glob(search_pattern, recursive=True) |
|
|
|
return [ |
|
datasets.SplitGenerator( |
|
name="full", |
|
gen_kwargs={"filepath": all_tsv_splits}, |
|
) |
|
] |
|
|
|
def _generate_examples(self, filepath): |
|
idx = 0 |
|
for file in filepath: |
|
yield idx, { |
|
'path': file |
|
} |
|
idx += 1 |