| import datasets |
| import os |
| import tarfile |
| import shutil |
| import subprocess |
| import tempfile |
|
|
| _VERSION = datasets.Version("1.0.0") |
|
|
| _URLS = { |
| "copydays_original": { |
| "images": [ |
| "https://dl.fbaipublicfiles.com/vissl/datasets/copydays_original.tar.gz" |
| ], |
| }, |
| "copydays_strong": { |
| "images": [ |
| "https://dl.fbaipublicfiles.com/vissl/datasets/copydays_strong.tar.gz" |
| ], |
| }, |
| } |
|
|
| _DESCRIPTION = ( |
| "Copydays dataset for copy detection and near-duplicate image retrieval evaluation." |
| ) |
|
|
| _CITATION = """\ |
| @inproceedings{jegou2008hamming, |
| title={Hamming embedding and weak geometric consistency for large scale image search}, |
| author={Jegou, Herve and Douze, Matthijs and Schmid, Cordelia}, |
| booktitle={European conference on computer vision}, |
| pages={304--317}, |
| year={2008}, |
| organization={Springer} |
| } |
| """ |
|
|
| BUILDER_CONFIGS = [ |
| datasets.BuilderConfig( |
| name="database", |
| version=_VERSION, |
| description="Copydays original split for copy detection evaluation. Original, unmodified images.", |
| ), |
| datasets.BuilderConfig( |
| name="query", |
| version=_VERSION, |
| description="Copydays query split for copy detection evaluation. Currently only contains the strong modifications.", |
| ), |
| ] |
|
|
|
|
| class Copydays(datasets.GeneratorBasedBuilder): |
| """Copydays copy detection dataset.""" |
|
|
| BUILDER_CONFIGS = BUILDER_CONFIGS |
| DEFAULT_CONFIG_NAME = "database" |
|
|
| def _download_and_extract(self, urls, cache_dir): |
| """Download archives using wget and extract them.""" |
| os.makedirs(cache_dir, exist_ok=True) |
|
|
| existing_files = [f for f in os.listdir(cache_dir) if f.endswith(".jpg")] |
| has_original = any(f.endswith("00") for f in existing_files) |
| has_strong = any( |
| not f.endswith("00") for f in existing_files if f.endswith(".jpg") |
| ) |
|
|
| if has_original and has_strong: |
| print( |
| f"Found existing extracted files in {cache_dir}, skipping download..." |
| ) |
| return [cache_dir] |
|
|
| for url in urls: |
| filename = url.split("/")[-1] |
| archive_path = os.path.join(cache_dir, filename) |
|
|
| |
| if not os.path.exists(archive_path): |
| print(f"Downloading {url}...") |
| result = subprocess.run( |
| ["wget", url, "-O", archive_path], capture_output=True, text=True |
| ) |
| if result.returncode != 0: |
| raise RuntimeError(f"Failed to download {url}: {result.stderr}") |
|
|
| marker_file = os.path.join(cache_dir, f".{filename}.extracted") |
| if not os.path.exists(marker_file): |
| print(f"Extracting {archive_path}...") |
| with tarfile.open(archive_path, "r:gz") as tar: |
| tar.extractall(cache_dir) |
| with open(marker_file, "w") as f: |
| f.write("extracted") |
|
|
| return [cache_dir] |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features( |
| { |
| "image": datasets.Image(), |
| "filename": datasets.Value( |
| "string" |
| ), |
| "split_type": datasets.Value("string"), |
| "block_id": datasets.Value( |
| "int32" |
| ), |
| "query_id": datasets.Value( |
| "int32" |
| ), |
| |
| } |
| ), |
| supervised_keys=None, |
| homepage="https://thoth.inrialpes.fr/~jegou/data.php.html#copydays", |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| |
| all_urls = [] |
| for dataset_type in _URLS.values(): |
| all_urls.extend(dataset_type["images"]) |
|
|
| cache_dir = tempfile.mkdtemp(prefix="copydays_") |
|
|
| try: |
| |
| archive_paths = dl_manager.download(all_urls) |
| extracted_paths = dl_manager.extract(archive_paths) |
|
|
| |
| if not isinstance(extracted_paths, list): |
| extracted_paths = [extracted_paths] |
| except Exception as e: |
| |
| print(f"HF download failed: {e}") |
| print( |
| "Falling back to wget download strategy... This typically works better for this dataset." |
| ) |
| extracted_paths = self._download_and_extract(all_urls, cache_dir) |
|
|
| return [ |
| datasets.SplitGenerator( |
| name="queries", |
| gen_kwargs={ |
| "image_dirs": extracted_paths, |
| "split_type": "queries", |
| "config_name": self.config.name, |
| }, |
| ), |
| datasets.SplitGenerator( |
| name="database", |
| gen_kwargs={ |
| "image_dirs": extracted_paths, |
| "split_type": "database", |
| "config_name": self.config.name, |
| }, |
| ), |
| ] |
|
|
| def _generate_examples(self, image_dirs, split_type, config_name): |
| """Generate examples for the dataset.""" |
| idx = 0 |
|
|
| for image_dir in image_dirs: |
| for root, dirs, files in os.walk(image_dir): |
| for file in files: |
| if file.lower().endswith((".jpg", ".jpeg", ".png", ".bmp", ".gif")): |
| file_path = os.path.join(root, file) |
| filename = file |
|
|
| |
| base_name = os.path.splitext(filename)[0] |
| if not base_name.isdigit() or len(base_name) != 6: |
| continue |
|
|
| block_id = int(base_name[:4]) |
| query_id_str = base_name[4:6] |
|
|
| if query_id_str != "00": |
| if split_type == "queries": |
| query_id = int(query_id_str) |
| actual_split_type = "strong" |
| yield idx, { |
| "image": file_path, |
| "filename": filename, |
| "split_type": actual_split_type, |
| "block_id": block_id, |
| "query_id": query_id, |
| } |
| idx += 1 |
| else: |
| actual_split_type = "original" |
| if split_type == "queries": |
| query_id = 0 |
| else: |
| query_id = -1 |
|
|
| yield idx, { |
| "image": file_path, |
| "filename": filename, |
| "split_type": actual_split_type, |
| "block_id": block_id, |
| "query_id": query_id, |
| } |
| idx += 1 |
|
|