| |
| import argparse |
| import os |
| from functools import partial |
| from multiprocessing import Pool |
| from pathlib import Path |
|
|
| import numpy as np |
| import pandas as pd |
| import requests |
| import utils |
| from PIL import Image |
| from tqdm import tqdm |
|
|
| METADATA_FILE = "published_images.csv" |
| METADATA_URL = "https://raw.githubusercontent.com/NationalGalleryOfArt/opendata/refs/heads/main/data" |
| IMG_URL = "https://api.nga.gov/iiif/%s/full/%s/0/default.jpg" |
| METADATA_FOLDER = "metadata" |
| EXTENSION = ".jpg" |
|
|
|
|
| def download_metadata(annotation_folder): |
| output_folder = annotation_folder / METADATA_FOLDER |
| output_folder.mkdir(exist_ok=True) |
| url = f"{METADATA_URL}/{METADATA_FILE}" |
| print(url) |
| response = requests.get(url) |
| if response.status_code == 200: |
| with open(output_folder / METADATA_FILE, "wb") as f: |
| f.write(response.content) |
|
|
|
|
| def download_url(row): |
| if np.isnan(row.maxpixels) or ( |
| row.maxpixels > row.width and row.maxpixels > row.height |
| ): |
| url = IMG_URL % (row.uuid, "full") |
| else: |
| url = IMG_URL % (row.uuid, f"!{row.maxpixels},{row.maxpixels}") |
| return url |
|
|
|
|
| def download_item(item, output_folder): |
| uuid, url = item |
| try: |
| if (output_folder / f"{uuid}{EXTENSION}").exists(): |
| print("skipping", uuid, "already downloaded") |
| return |
| response = requests.get(url) |
| if response.status_code == 200: |
| with open(output_folder / f"{uuid}{EXTENSION}", "wb") as f: |
| f.write(response.content) |
| except: |
| print("errored", item) |
| return |
|
|
|
|
| def remove_non_compliant_image(item, output_folder): |
| uuid, max_pixels = item |
| if np.isnan(max_pixels): |
| return |
| if not (output_folder / f"{uuid}{EXTENSION}").exists(): |
| return |
| img = Image.open(output_folder / f"{uuid}{EXTENSION}") |
| if img.width > max_pixels or img.height > max_pixels: |
| os.remove(output_folder / f"{uuid}{EXTENSION}") |
| return uuid |
|
|
|
|
| def reshape_image(rel_path, filename_size_map, output_folder): |
| w, h = filename_size_map[rel_path] |
| path = output_folder / f"{rel_path}" |
| img = Image.open(path) |
| if img.width != w or img.height != h: |
| new_size = (w, h) |
| resized_img = img.resize(new_size) |
| resized_img.save(path) |
|
|
|
|
| def main(args, workers=20): |
| raw_folder = Path(args.raw_images_folder) |
| processed_folder = Path(args.processed_images_folder) |
| utils.setup(raw_folder) |
| utils.setup(processed_folder) |
| uuids = utils.get_image_ids(args.annotation_file) |
| filename_size_map = utils.get_filename_size_map(args.annotation_file) |
| if not ((raw_folder / METADATA_FOLDER) / METADATA_FILE).exists(): |
| download_metadata(raw_folder) |
|
|
| metadata = pd.read_csv((raw_folder / METADATA_FOLDER) / METADATA_FILE) |
| metadata["download_url"] = metadata.apply(download_url, axis=1) |
| available_uuids = list(uuids.intersection(set(metadata["uuid"].tolist()))) |
| print(len(available_uuids), "available for download out of", len(uuids), "target") |
| url_data = list( |
| metadata.set_index("uuid") |
| .loc[available_uuids] |
| .to_dict()["download_url"] |
| .items() |
| ) |
|
|
| download_single = partial(download_item, output_folder=(processed_folder)) |
|
|
| print("Preparing to download", len(url_data), "items") |
| with Pool(20) as p: |
| for _ in tqdm(p.imap(download_single, url_data), total=len(url_data)): |
| continue |
| check_img_size = partial( |
| remove_non_compliant_image, output_folder=(processed_folder) |
| ) |
| max_pixels_dict_all = metadata.set_index("uuid").to_dict()["maxpixels"] |
| max_pixels_dict = {item[0]: max_pixels_dict_all[item[0]] for item in url_data} |
| print("Checking all images within size constraints") |
| non_compliant = set() |
| with Pool(20) as p: |
| for each in tqdm( |
| p.imap(check_img_size, max_pixels_dict.items()), total=len(max_pixels_dict) |
| ): |
| if each is not None: |
| non_compliant.add(each) |
| print(len(non_compliant), "not compliant size, removed") |
|
|
| reshape_single = partial( |
| reshape_image, |
| filename_size_map=(filename_size_map), |
| output_folder=(processed_folder), |
| ) |
| rel_paths = os.listdir(args.processed_images_folder) |
| print("Preparing to reshape", len(rel_paths), "items") |
| with Pool(20) as p: |
| for _ in tqdm(p.imap(reshape_single, rel_paths), total=len(rel_paths)): |
| continue |
|
|
|
|
| if __name__ == "__main__": |
| parser = argparse.ArgumentParser() |
| parser.add_argument("--annotation_file", help="Path to annotation file") |
| parser.add_argument("--raw_images_folder", help="Path to downloaded images") |
| parser.add_argument("--processed_images_folder", help="Path to processed images") |
| args = parser.parse_args() |
| main(args) |
|
|