| import argparse |
| import json |
| from pathlib import Path |
|
|
| import pandas as pd |
|
|
| from utils import fuse_csv, fuse_tif |
|
|
| if __name__ == "__main__": |
| parser = argparse.ArgumentParser() |
| parser.add_argument("--checkpoint", type=str, required=True, |
| help="Checkpoint config") |
| parser.add_argument("--last-only", action="store_true", |
| help="Only do the last dataset (for debugging)") |
| parser.add_argument("--recover_unfinished", action="store_true", |
| help="Recover from samples_unfinished.json and skip finished datasets") |
| args = parser.parse_args() |
|
|
| |
| |
| |
| with open(f"./src/fuse_esdac/configs/{args.checkpoint}.json", "r") as f: |
| cfg = json.load(f) |
|
|
| input_json = Path(f"./src/fuse_esdac/outputs/{cfg.get('input')}/samples.json") |
| datasets = cfg.get("datasets", []) |
| if args.last_only and datasets: |
| datasets = datasets[-1:] |
|
|
| |
| |
| |
| out_dir = Path(f"./src/fuse_esdac/outputs/{args.checkpoint}") |
| out_dir.mkdir(parents=True, exist_ok=True) |
|
|
| out_json = out_dir / f"samples.json" |
| out_var_csv = out_dir / f"stat_variable.csv" |
| unfinished_json = out_dir / "samples_unfinished.json" |
| finished_datasets_json = out_dir / "datasets_finished.json" |
|
|
| |
| |
| |
| if args.recover_unfinished: |
| if not unfinished_json.exists(): |
| raise FileNotFoundError(f"--recover_unfinished requires {unfinished_json} to exist") |
| with open(unfinished_json, "r") as f: |
| big_output_dict = json.load(f) |
|
|
| if finished_datasets_json.exists(): |
| with open(finished_datasets_json, "r") as f: |
| finished_datasets = set(json.load(f)) |
| else: |
| finished_datasets = set() |
|
|
| |
| datasets = [d for d in datasets if d not in finished_datasets] |
| else: |
| finished_datasets = set() |
| if input_json and Path(input_json).exists(): |
| with open(input_json, "r") as f: |
| big_output_dict = json.load(f) |
| else: |
| big_output_dict = {} |
|
|
| |
| |
| |
| for dataset in datasets: |
| try: |
| data_dir = Path(f"./datasets/esdac/{dataset}/processed") |
| schema_path = Path(f"./src/esdac/{dataset}/fuse_schema.json") |
| with open(schema_path, "r") as f: |
| schema = json.load(f) |
|
|
| for filename in schema.keys(): |
| if filename.endswith(".csv"): |
| csv_path = data_dir / filename |
| big_output_dict = fuse_csv( |
| source_file=csv_path, |
| schema=schema[filename], |
| source_name=dataset, |
| current_dict=big_output_dict, |
| ) |
| elif filename.endswith("@tif"): |
| tif_path = data_dir / filename[:-4] |
| big_output_dict = fuse_tif( |
| source_file=tif_path, |
| schema=schema[filename], |
| source_name=dataset, |
| current_dict=big_output_dict, |
| ) |
|
|
| |
| finished_datasets.add(dataset) |
| with open(finished_datasets_json, "w") as f: |
| json.dump(sorted(finished_datasets), f, indent=2, ensure_ascii=False) |
|
|
| except Exception: |
| |
| with open(unfinished_json, "w") as f: |
| json.dump(big_output_dict, f, indent=2, ensure_ascii=False) |
| raise |
|
|
| |
| |
| |
| with open(out_json, "w") as f: |
| json.dump(big_output_dict, f, indent=2, ensure_ascii=False) |
|
|
| |
| |
| |
| table = {} |
| for sample_key, sample in big_output_dict.items(): |
| prefix = sample_key.split("_")[0] |
| if prefix not in table: |
| table[prefix] = {} |
| for domain, section in sample.items(): |
| if isinstance(section, dict): |
| for var, leaf in section.items(): |
| if isinstance(leaf, dict): |
| if leaf.get("value") not in (None, "", "null"): |
| key = f"{domain}:{var}" |
| table[prefix][key] = table[prefix].get(key, 0) + 1 |
|
|
| |
| sample_count = {} |
| for sample_key in big_output_dict.keys(): |
| prefix = sample_key.split("_")[0] |
| sample_count[prefix] = sample_count.get(prefix, 0) + 1 |
|
|
| df_var = ( |
| pd.DataFrame.from_dict(table, orient="index") |
| .fillna(0) |
| .infer_objects() |
| .astype(int) |
| ) |
|
|
| |
| all_prefixes = sorted(sample_count.keys()) |
| if not df_var.empty: |
| df_var = df_var.reindex(all_prefixes, fill_value=0) |
| else: |
| df_var = pd.DataFrame(index=all_prefixes) |
|
|
| |
| df_var["TOTAL_SAMPLES"] = df_var.index.map(lambda p: sample_count.get(p, 0)).astype(int) |
|
|
| |
| totals_no_samples_col = df_var.drop(columns=["TOTAL_SAMPLES"]).sum(axis=0).astype(int) |
| df_var.loc["SUM"] = 0 |
| df_var.loc["SUM", totals_no_samples_col.index] = totals_no_samples_col |
| df_var.loc["SUM", "TOTAL_SAMPLES"] = int(sum(sample_count.values())) |
| df_var = df_var.astype(int) |
|
|
| df_var.to_csv(out_var_csv) |
|
|