danish-dynaword / src /dynaword /update_descriptive_statistics.py
Kenneth Enevoldsen
update desc stats
d36009a unverified
raw
history blame
5.05 kB
"""
A simple CLI to updates descriptive statistics on all datasets.
Example use:
uv run src/dynaword/update_descriptive_statistics.py --dataset wikisource
"""
import argparse
import json
import logging
from pathlib import Path
from typing import cast
import plotly.express as px
from datasets import Dataset, load_dataset
from dynaword.datasheet import DataSheet
from dynaword.descriptive_stats import DescriptiveStatsOverview
from dynaword.git_utilities import (
check_is_ancestor,
get_latest_revision,
)
from dynaword.paths import repo_path
from dynaword.tables import create_overview_table, create_overview_table_str
logger = logging.getLogger(__name__)
main_sheet = DataSheet.load_from_path(repo_path / "README.md")
_datasets = [
cfg["config_name"] # type: ignore
for cfg in main_sheet.frontmatter["configs"] # type: ignore
if cfg["config_name"] != "default" # type: ignore
]
logger = logging.getLogger(__name__)
def create_domain_distribution_plot(
save_dir: Path = repo_path,
):
df = create_overview_table(
add_readable_tokens=False, add_total_row=False, add_readme_references=False
)
fig = px.sunburst(df, path=["Domain", "Source"], values="N. Tokens")
fig.update_traces(textinfo="label+percent entry")
fig.update_layout(title="Dataset Distribution by Domain and Source")
img_path = save_dir / "images"
img_path.mkdir(parents=False, exist_ok=True)
save_path = img_path / "domain_distribution.png"
fig.write_image(
save_path,
width=800,
height=800,
scale=2,
)
def update_dataset(
dataset_name: str,
force: bool = False,
) -> None:
dataset_path = (
repo_path / "data" / dataset_name if dataset_name != "default" else repo_path
)
if dataset_name == "default":
readme_name = "README.md"
else:
readme_name = f"{dataset_name}.md"
rev = get_latest_revision(dataset_path)
desc_stats_path = dataset_path / "descriptive_stats.json"
markdown_path = dataset_path / readme_name
if desc_stats_path.exists() and force is False:
with desc_stats_path.open("r") as f:
last_update = json.load(f).get("revision", None)
if last_update is None:
logger.warning(f"revision is not defined in {desc_stats_path}.")
elif check_is_ancestor(ancestor_rev=last_update, rev=rev):
logger.info(
f"descriptive statistics for '{dataset_name}' is already up to date, skipping."
)
return
logger.info(f"Computing descriptive stats for: {dataset_name}")
ds = load_dataset(str(repo_path), dataset_name, split="train")
ds = cast(Dataset, ds)
desc_stats = DescriptiveStatsOverview.from_dataset(ds)
desc_stats.to_disk(desc_stats_path)
logger.info(f"Updating datasheet for: {dataset_name}")
sheet = DataSheet.load_from_path(markdown_path)
sheet.body = sheet.add_descriptive_stats(descriptive_stats=desc_stats)
sheet.body = sheet.add_sample_and_description(ds)
sheet.body = sheet.add_dataset_plots(ds, create_plot=True)
if dataset_name == "default":
logger.info("Updating Overview table")
package = create_overview_table_str()
sheet.body = sheet.replace_tag(package=package, tag="MAIN TABLE")
create_domain_distribution_plot()
sheet.write_to_path()
def create_parser():
parser = argparse.ArgumentParser(
description="Calculated descriptive statistics of the datasets in tha data folder"
)
parser.add_argument(
"--dataset",
default=None,
type=str,
help="Use to specify if you only want to compute the statistics from a singular dataset.",
)
parser.add_argument(
"--logging_level",
default=20,
type=int,
help="Sets the logging level. Default to 20 (INFO), other reasonable levels are 10 (DEBUG) and 30 (WARNING).",
)
parser.add_argument(
"--force",
type=bool,
default=False,
action=argparse.BooleanOptionalAction,
help="Should the statistics be forcefully recomputed. By default it checks the difference in commit ids.",
)
parser.add_argument(
"--repo_path",
default=str(repo_path),
type=str,
help="The repository where to calculate the descriptive statistics from",
)
return parser
def main(
dataset: str | None = None,
logging_level: int = 20,
force: bool = False,
repo_path: Path = repo_path,
) -> None:
logging.basicConfig(level=logging_level)
if dataset:
update_dataset(dataset, force=force)
else:
for dataset_name in _datasets:
update_dataset(dataset_name, force=force)
update_dataset("default", force=force)
if __name__ == "__main__":
parser = create_parser()
args = parser.parse_args()
main(
args.dataset,
logging_level=args.logging_level,
force=args.force,
repo_path=Path(args.repo_path),
)