File size: 11,583 Bytes
9bc98d9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
import json
import os
import pickle
from collections import defaultdict, OrderedDict

import numpy as np
import pandas as pd
from tqdm import tqdm

# ----------------------------- CONFIG -------------------------------- #
INPUT_JSON = "./src/fuse_esdac/outputs/03_photo_gadm/samples_with_photo_gadm.pkl"
OUTPUT_CSV = "./src/fuse_esdac/outputs/04_table/final_table.csv"
OUTPUT_DICT_PKL = "./src/fuse_esdac/outputs/04_table/final_dict.pkl"
OUTPUT_DICT_JSON = "./src/fuse_esdac/outputs/04_table/final_dict.json"
OUTPUT_META = "./src/fuse_esdac/outputs/04_table/column_meta.json"
OUTPUT_META_NAMES = "./src/fuse_esdac/outputs/04_table/column_names.json"
# --------------------------------------------------------------------- #

# Null normalization rules
NULL_STRINGS = {
    "nan", "null", "none", "", "na", "n/a", "missing", "unknown", "unknown1"
}


def is_null_value(v):
    # Real null
    if v is None:
        return True
    # String forms
    if isinstance(v, str):
        s = v.strip().lower()
        if s in NULL_STRINGS:
            return True
    return False


FILL_VALUES_TO_NULL = {-9000, -90000}
NEGATIVE_FEATURE_KEYWORDS = ("density", "percentage", "content", "stock")
NONE_POSITIVE_FEATURE_KEYWORDS = ("ph_",)


def numeric_or_none(v):
    try:
        return float(v)
    except (TypeError, ValueError):
        return None


def clean_numeric_scalar_value(header, v):
    if is_null_value(v):
        return None

    x = numeric_or_none(v)
    if x is None:
        return v

    if x in FILL_VALUES_TO_NULL:
        return None

    h = header.lower()
    if any(k in h for k in NEGATIVE_FEATURE_KEYWORDS) and x < -0.00001:
        return None
    if any(k in h for k in NONE_POSITIVE_FEATURE_KEYWORDS) and x < 0.00001:
        return None

    return v


def clean_numeric_vector_value(header, v):
    if is_null_value(v):
        return None

    if not isinstance(v, list):
        return v

    cleaned = [
        clean_numeric_scalar_value(header, x)
        for x in v
    ]

    if all(x is None for x in cleaned):
        return None

    return cleaned


def main():
    with open(INPUT_JSON, "rb") as f:
        data = pickle.load(f)

    # 1) Discover variables and enforce unit consistency on (theme:var)
    var_to_unit = {}  # "theme:var" -> unit_str ('' if missing)
    for sid, sample in data.items():
        # force presence of required non-variable fields (surface schema issues early)
        _ = sample["LAT_LONG"]
        _ = sample["GADM_IDS"]
        _ = sample["GADM_NAMES"]
        _ = sample["COUNTRY_CODE"]

        for theme, theme_obj in sample.items():
            if theme in ("LAT_LONG", "GADM_IDS", "GADM_NAMES", "COUNTRY_CODE",
                         "SAMPLE_DATE", "SAMPLE_DEPTH_RANGE_CM", "SAMPLE_SOURCE_DATASET"):
                continue

            # theme_obj is expected to be a dict of variables
            for var, var_obj in theme_obj.items():
                # required fields per your spec (value, source_dataset, distance_to_grid; unit optional)
                _ = var_obj["value"]
                _ = var_obj["source_dataset"]
                _ = var_obj["distance_to_grid (m)"]
                unit = var_obj.get("unit", "")  # empty string if missing

                key = f"{theme}:{var}"
                if key not in var_to_unit:
                    var_to_unit[key] = unit
                else:
                    if var_to_unit[key] != unit:
                        print(f"WARNING: inconsistent unit for '{key}': "
                              f"using '{var_to_unit[key]}' but encountered '{unit}'")

    # 2) Build a stable mapping to final column headers with unit parentheses (including empty ())
    def header_for(key):
        unit = var_to_unit[key]
        return f"{key} ({unit})" if unit else key

    base_cols = ["id", "LAT_LONG", "GADM_IDS", "GADM_NAMES", "COUNTRY_CODE",
                 "SAMPLE_DATE", "SAMPLE_DEPTH_RANGE_CM", "SAMPLE_SOURCE_DATASET"]
    var_cols = [header_for(key) for key in var_to_unit.keys()]
    all_cols = base_cols + var_cols

    # 3) Populate rows and collect meta
    meta_sources = defaultdict(set)  # header -> set of datasets
    meta_distances = defaultdict(list)  # header -> list of numeric distances (non-null only)

    table = []
    for sid, sample in tqdm(data.items(), desc="Building table"):
        row = OrderedDict.fromkeys(all_cols, None)

        row["id"] = sid
        row["LAT_LONG"] = sample["LAT_LONG"]
        row["GADM_IDS"] = sample["GADM_IDS"]
        row["GADM_NAMES"] = sample["GADM_NAMES"]
        row["COUNTRY_CODE"] = sample["COUNTRY_CODE"]
        row["SAMPLE_DATE"] = sample.get("SAMPLE_DATE")
        row["SAMPLE_DEPTH_RANGE_CM"] = sample.get("SAMPLE_DEPTH_RANGE_CM")
        row["SAMPLE_SOURCE_DATASET"] = sample.get("SAMPLE_SOURCE_DATASET")

        for theme, theme_obj in sample.items():
            if theme in ("LAT_LONG", "GADM_IDS", "GADM_NAMES", "COUNTRY_CODE",
                         "SAMPLE_DATE", "SAMPLE_DEPTH_RANGE_CM", "SAMPLE_SOURCE_DATASET"):
                continue

            for var, var_obj in theme_obj.items():
                key = f"{theme}:{var}"
                header = header_for(key)

                v = var_obj["value"]
                if is_null_value(v):
                    v = None
                row[header] = v  # write to final header (with unit parentheses)

                # --- meta: collect sources regardless of v ---
                src = var_obj["source_dataset"]
                if isinstance(src, str) and src.strip():
                    meta_sources[header].add(src)

                # --- meta: collect non-null distances only ---
                dist = var_obj["distance_to_grid (m)"]
                if not is_null_value(dist):
                    try:
                        meta_distances[header].append(float(dist))
                    except (TypeError, ValueError):
                        # Surface bad distance type immediately
                        raise TypeError(f"distance_to_grid (m) not numeric for '{header}' in sample '{sid}': {dist!r}")

        table.append(row)

    df = pd.DataFrame(table, columns=all_cols)

    # 4) Warn for columns that are entirely null
    for col in var_cols:
        if df[col].isna().all():
            print(f"WARNING: column has all null values -> {col}")

    # 5) Build column_meta.json before cleaning
    def py_type(x):
        if isinstance(x, bool):
            return "bool"
        if isinstance(x, int):
            return "int"
        if isinstance(x, float):
            return "float"
        if isinstance(x, str):
            return "string"
        return None  # ignore other types

    meta = {}
    total_rows = len(df)

    for header in var_cols:
        # ----- distance stats -----
        dists = np.array(meta_distances[header], dtype=float) if meta_distances[header] else np.array([])
        if dists.size > 0:
            stats = {
                "min": float(np.min(dists)),
                "max": float(np.max(dists)),
                "mean": float(np.mean(dists)),
                "std": float(np.std(dists)),
                "median": float(np.median(dists)),
            }
        else:
            stats = {
                "min": None,
                "max": None,
                "mean": None,
                "std": None,
                "median": None,
            }

        # ----- detect array-valued or scalar -----
        col_vals = df[header].tolist()
        first_non_null = next((v for v in col_vals if v is not None), None)

        if isinstance(first_non_null, list):
            is_array = True
            flat_vals = []
            for v in col_vals:
                if isinstance(v, list):
                    flat_vals.extend(v)
            type_counts = {}
            for v in flat_vals:
                t = py_type(v)
                if t:
                    type_counts[t] = type_counts.get(t, 0) + 1
        else:
            is_array = False
            type_counts = {}
            for v in col_vals:
                t = py_type(v)
                if t:
                    type_counts[t] = type_counts.get(t, 0) + 1

        datatype = max(type_counts.items(), key=lambda x: x[1])[0] if type_counts else "unknown"
        null_fraction = float(df[header].isna().sum()) / total_rows

        meta[header] = {
            "source_datasets": sorted(meta_sources[header]),
            "distance_to_grid_stats (m)": stats,
            "null_fraction": null_fraction,
            "datatype": datatype,
            "is_array_valued": is_array,
            "description": "",
        }

    # 6) Clean numeric scalar columns only
    cleaned_cells = {}
    cleaned_columns = defaultdict(int)

    for header in var_cols:
        if meta[header]["datatype"] not in ("int", "float"):
            continue
        if not meta[header]["is_array_valued"]:
            for row_i, v in df[header].items():
                new_v = clean_numeric_scalar_value(header, v)
                if not (
                        pd.isna(new_v) and pd.isna(v)
                ) and new_v != v:
                    df.at[row_i, header] = new_v
                    cleaned_cells[(row_i, header)] = new_v
                    cleaned_columns[header] += 1
        else:
            for row_i, v in df[header].items():
                new_v = clean_numeric_vector_value(header, v)
                if new_v != v:
                    df.at[row_i, header] = new_v
                    cleaned_cells[(row_i, header)] = new_v
                    cleaned_columns[header] += 1

    if cleaned_columns:
        print("Cleaned columns:")
        for header, n in sorted(cleaned_columns.items()):
            print(f"  {header}: {n} cells")
    else:
        print("Cleaned columns: none")

    # 7) Write cleaned values back to dict
    for row_i, row in df.iterrows():
        sid = row["id"]
        sample = data[sid]

        for theme, theme_obj in sample.items():
            if theme in ("LAT_LONG", "GADM_IDS", "GADM_NAMES", "COUNTRY_CODE",
                         "SAMPLE_DATE", "SAMPLE_DEPTH_RANGE_CM", "SAMPLE_SOURCE_DATASET"):
                continue

            for var, var_obj in theme_obj.items():
                key = f"{theme}:{var}"
                header = header_for(key)

                if (row_i, header) in cleaned_cells:
                    var_obj["value"] = cleaned_cells[(row_i, header)]

    # 8) Save outputs
    os.makedirs(os.path.dirname(OUTPUT_CSV), exist_ok=True)

    df_out = df.copy()
    df_out = df_out.replace({np.nan: None})
    df_out.to_csv(OUTPUT_CSV, index=False, encoding="utf-8", na_rep="")

    # Write cleaned dict as pickle
    with open(OUTPUT_DICT_PKL, "wb") as f:
        pickle.dump(data, f)

    # Write cleaned dict as json
    with open(OUTPUT_DICT_JSON, "w", encoding="utf-8") as f:
        json.dump(
            data,
            f,
            indent=2,
            ensure_ascii=False
        )

    with open(OUTPUT_META, "w", encoding="utf-8") as f:
        json.dump(meta, f, indent=2)

    with open(OUTPUT_META_NAMES, "w", encoding="utf-8") as f:
        json.dump(
            {"column_names": sorted(meta.keys())},
            f,
            indent=2,
            ensure_ascii=False,
        )

    print("Done.")
    print(f"Rows: {len(df)}  Columns: {len(df.columns)}")
    print(f"Cleaned cells: {len(cleaned_cells)}")
    print(
        f"Saved: {OUTPUT_CSV}, "
        f"{OUTPUT_DICT_PKL}, {OUTPUT_DICT_JSON}, "
        f"{OUTPUT_META}, {OUTPUT_META_NAMES}"
    )


if __name__ == "__main__":
    main()