| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| import hashlib |
| import json |
| import logging |
| import os |
| import shutil |
| import sys |
| import tarfile |
| import tempfile |
| import warnings |
| import zipfile |
| from pathlib import Path |
| from typing import TYPE_CHECKING, Any |
| from urllib.error import ContentTooShortError, HTTPError, URLError |
| from urllib.parse import urlparse |
| from urllib.request import urlopen, urlretrieve |
|
|
| from monai.config.type_definitions import PathLike |
| from monai.utils import look_up_option, min_version, optional_import |
|
|
| gdown, has_gdown = optional_import("gdown", "4.7.3") |
|
|
| if TYPE_CHECKING: |
| from tqdm import tqdm |
|
|
| has_tqdm = True |
| else: |
| tqdm, has_tqdm = optional_import("tqdm", "4.47.0", min_version, "tqdm") |
|
|
| __all__ = ["check_hash", "download_url", "extractall", "download_and_extract", "get_logger", "SUPPORTED_HASH_TYPES"] |
|
|
| DEFAULT_FMT = "%(asctime)s - %(levelname)s - %(message)s" |
| SUPPORTED_HASH_TYPES = {"md5": hashlib.md5, "sha1": hashlib.sha1, "sha256": hashlib.sha256, "sha512": hashlib.sha512} |
|
|
|
|
| def get_logger( |
| module_name: str = "monai.apps", |
| fmt: str = DEFAULT_FMT, |
| datefmt: str | None = None, |
| logger_handler: logging.Handler | None = None, |
| ) -> logging.Logger: |
| """ |
| Get a `module_name` logger with the specified format and date format. |
| By default, the logger will print to `stdout` at the INFO level. |
| If `module_name` is `None`, return the root logger. |
| `fmt` and `datafmt` are passed to a `logging.Formatter` object |
| (https://docs.python.org/3/library/logging.html#formatter-objects). |
| `logger_handler` can be used to add an additional handler. |
| """ |
| adds_stdout_handler = module_name is not None and module_name not in logging.root.manager.loggerDict |
| logger = logging.getLogger(module_name) |
| logger.propagate = False |
| logger.setLevel(logging.INFO) |
| if adds_stdout_handler: |
| handler = logging.StreamHandler(sys.stdout) |
| formatter = logging.Formatter(fmt=fmt, datefmt=datefmt) |
| handler.setFormatter(formatter) |
| logger.addHandler(handler) |
| if logger_handler is not None: |
| logger.addHandler(logger_handler) |
| return logger |
|
|
|
|
| |
| logger = get_logger("monai.apps") |
| __all__.append("logger") |
|
|
|
|
| def _basename(p: PathLike) -> str: |
| """get the last part of the path (removing the trailing slash if it exists)""" |
| sep = os.path.sep + (os.path.altsep or "") + "/ " |
| return Path(f"{p}".rstrip(sep)).name |
|
|
|
|
| def _download_with_progress(url: str, filepath: Path, progress: bool = True) -> None: |
| """ |
| Retrieve file from `url` to `filepath`, optionally showing a progress bar. |
| """ |
| try: |
| if has_tqdm and progress: |
|
|
| class TqdmUpTo(tqdm): |
| """ |
| Provides `update_to(n)` which uses `tqdm.update(delta_n)`. |
| Inspired by the example in https://github.com/tqdm/tqdm. |
| """ |
|
|
| def update_to(self, b: int = 1, bsize: int = 1, tsize: int | None = None) -> None: |
| """ |
| Args: |
| b: number of blocks transferred so far, default: 1. |
| bsize: size of each block (in tqdm units), default: 1. |
| tsize: total size (in tqdm units). if None, remains unchanged. |
| """ |
| if tsize is not None: |
| self.total = tsize |
| self.update(b * bsize - self.n) |
|
|
| with TqdmUpTo(unit="B", unit_scale=True, unit_divisor=1024, miniters=1, desc=_basename(filepath)) as t: |
| urlretrieve(url, filepath, reporthook=t.update_to) |
| else: |
| if not has_tqdm and progress: |
| warnings.warn("tqdm is not installed, will not show the downloading progress bar.") |
| urlretrieve(url, filepath) |
| except (URLError, HTTPError, ContentTooShortError, OSError) as e: |
| logger.error(f"Download failed from {url} to {filepath}.") |
| raise e |
|
|
|
|
| def check_hash(filepath: PathLike, val: str | None = None, hash_type: str = "md5") -> bool: |
| """ |
| Verify hash signature of specified file. |
| |
| Args: |
| filepath: path of source file to verify hash value. |
| val: expected hash value of the file. |
| hash_type: type of hash algorithm to use, default is `"md5"`. |
| The supported hash types are `"md5"`, `"sha1"`, `"sha256"`, `"sha512"`. |
| See also: :py:data:`monai.apps.utils.SUPPORTED_HASH_TYPES`. |
| |
| """ |
| if val is None: |
| logger.info(f"Expected {hash_type} is None, skip {hash_type} check for file {filepath}.") |
| return True |
| actual_hash_func = look_up_option(hash_type.lower(), SUPPORTED_HASH_TYPES) |
|
|
| if sys.version_info >= (3, 9): |
| actual_hash = actual_hash_func(usedforsecurity=False) |
| else: |
| actual_hash = actual_hash_func() |
|
|
| try: |
| with open(filepath, "rb") as f: |
| for chunk in iter(lambda: f.read(1024 * 1024), b""): |
| actual_hash.update(chunk) |
| except Exception as e: |
| logger.error(f"Exception in check_hash: {e}") |
| return False |
| if val != actual_hash.hexdigest(): |
| logger.error(f"check_hash failed {actual_hash.hexdigest()}.") |
| return False |
|
|
| logger.info(f"Verified '{_basename(filepath)}', {hash_type}: {val}.") |
| return True |
|
|
|
|
| def download_url( |
| url: str, |
| filepath: PathLike = "", |
| hash_val: str | None = None, |
| hash_type: str = "md5", |
| progress: bool = True, |
| **gdown_kwargs: Any, |
| ) -> None: |
| """ |
| Download file from specified URL link, support process bar and hash check. |
| |
| Args: |
| url: source URL link to download file. |
| filepath: target filepath to save the downloaded file (including the filename). |
| If undefined, `os.path.basename(url)` will be used. |
| hash_val: expected hash value to validate the downloaded file. |
| if None, skip hash validation. |
| hash_type: 'md5' or 'sha1', defaults to 'md5'. |
| progress: whether to display a progress bar. |
| gdown_kwargs: other args for `gdown` except for the `url`, `output` and `quiet`. |
| these args will only be used if download from google drive. |
| details of the args of it: |
| https://github.com/wkentaro/gdown/blob/main/gdown/download.py |
| |
| Raises: |
| RuntimeError: When the hash validation of the ``filepath`` existing file fails. |
| RuntimeError: When a network issue or denied permission prevents the |
| file download from ``url`` to ``filepath``. |
| URLError: See urllib.request.urlretrieve. |
| HTTPError: See urllib.request.urlretrieve. |
| ContentTooShortError: See urllib.request.urlretrieve. |
| IOError: See urllib.request.urlretrieve. |
| RuntimeError: When the hash validation of the ``url`` downloaded file fails. |
| |
| """ |
| if not filepath: |
| filepath = Path(".", _basename(url)).resolve() |
| logger.info(f"Default downloading to '{filepath}'") |
| filepath = Path(filepath) |
| if filepath.exists(): |
| if not check_hash(filepath, hash_val, hash_type): |
| raise RuntimeError( |
| f"{hash_type} check of existing file failed: filepath={filepath}, expected {hash_type}={hash_val}." |
| ) |
| logger.info(f"File exists: {filepath}, skipped downloading.") |
| return |
| try: |
| with tempfile.TemporaryDirectory() as tmp_dir: |
| tmp_name = Path(tmp_dir, _basename(filepath)) |
| if urlparse(url).netloc == "drive.google.com": |
| if not has_gdown: |
| raise RuntimeError("To download files from Google Drive, please install the gdown dependency.") |
| if "fuzzy" not in gdown_kwargs: |
| gdown_kwargs["fuzzy"] = True |
| gdown.download(url, f"{tmp_name}", quiet=not progress, **gdown_kwargs) |
| elif urlparse(url).netloc == "cloud-api.yandex.net": |
| with urlopen(url) as response: |
| code = response.getcode() |
| if code == 200: |
| download_url = json.load(response)["href"] |
| _download_with_progress(download_url, tmp_name, progress=progress) |
| else: |
| raise RuntimeError( |
| f"Download of file from {download_url}, received from {url} " |
| + f" to {filepath} failed due to network issue or denied permission." |
| ) |
| else: |
| _download_with_progress(url, tmp_name, progress=progress) |
| if not tmp_name.exists(): |
| raise RuntimeError( |
| f"Download of file from {url} to {filepath} failed due to network issue or denied permission." |
| ) |
| file_dir = filepath.parent |
| if file_dir: |
| os.makedirs(file_dir, exist_ok=True) |
| shutil.move(f"{tmp_name}", f"{filepath}") |
| except (PermissionError, NotADirectoryError): |
| pass |
| logger.info(f"Downloaded: {filepath}") |
| if not check_hash(filepath, hash_val, hash_type): |
| raise RuntimeError( |
| f"{hash_type} check of downloaded file failed: URL={url}, " |
| f"filepath={filepath}, expected {hash_type}={hash_val}." |
| ) |
|
|
|
|
| def extractall( |
| filepath: PathLike, |
| output_dir: PathLike = ".", |
| hash_val: str | None = None, |
| hash_type: str = "md5", |
| file_type: str = "", |
| has_base: bool = True, |
| ) -> None: |
| """ |
| Extract file to the output directory. |
| Expected file types are: `zip`, `tar.gz` and `tar`. |
| |
| Args: |
| filepath: the file path of compressed file. |
| output_dir: target directory to save extracted files. |
| hash_val: expected hash value to validate the compressed file. |
| if None, skip hash validation. |
| hash_type: 'md5' or 'sha1', defaults to 'md5'. |
| file_type: string of file type for decompressing. Leave it empty to infer the type from the filepath basename. |
| has_base: whether the extracted files have a base folder. This flag is used when checking if the existing |
| folder is a result of `extractall`, if it is, the extraction is skipped. For example, if A.zip is unzipped |
| to folder structure `A/*.png`, this flag should be True; if B.zip is unzipped to `*.png`, this flag should |
| be False. |
| |
| Raises: |
| RuntimeError: When the hash validation of the ``filepath`` compressed file fails. |
| NotImplementedError: When the ``filepath`` file extension is not one of [zip", "tar.gz", "tar"]. |
| |
| """ |
| if has_base: |
| |
| cache_dir = Path(output_dir, _basename(filepath).split(".")[0]) |
| else: |
| cache_dir = Path(output_dir) |
| if cache_dir.exists() and next(cache_dir.iterdir(), None) is not None: |
| logger.info(f"Non-empty folder exists in {cache_dir}, skipped extracting.") |
| return |
| filepath = Path(filepath) |
| if hash_val and not check_hash(filepath, hash_val, hash_type): |
| raise RuntimeError( |
| f"{hash_type} check of compressed file failed: " f"filepath={filepath}, expected {hash_type}={hash_val}." |
| ) |
| logger.info(f"Writing into directory: {output_dir}.") |
| _file_type = file_type.lower().strip() |
| if filepath.name.endswith("zip") or _file_type == "zip": |
| zip_file = zipfile.ZipFile(filepath) |
| zip_file.extractall(output_dir) |
| zip_file.close() |
| return |
| if filepath.name.endswith("tar") or filepath.name.endswith("tar.gz") or "tar" in _file_type: |
| tar_file = tarfile.open(filepath) |
| tar_file.extractall(output_dir) |
| tar_file.close() |
| return |
| raise NotImplementedError( |
| f'Unsupported file type, available options are: ["zip", "tar.gz", "tar"]. name={filepath} type={file_type}.' |
| ) |
|
|
|
|
| def download_and_extract( |
| url: str, |
| filepath: PathLike = "", |
| output_dir: PathLike = ".", |
| hash_val: str | None = None, |
| hash_type: str = "md5", |
| file_type: str = "", |
| has_base: bool = True, |
| progress: bool = True, |
| ) -> None: |
| """ |
| Download file from URL and extract it to the output directory. |
| |
| Args: |
| url: source URL link to download file. |
| filepath: the file path of the downloaded compressed file. |
| use this option to keep the directly downloaded compressed file, to avoid further repeated downloads. |
| output_dir: target directory to save extracted files. |
| default is the current directory. |
| hash_val: expected hash value to validate the downloaded file. |
| if None, skip hash validation. |
| hash_type: 'md5' or 'sha1', defaults to 'md5'. |
| file_type: string of file type for decompressing. Leave it empty to infer the type from url's base file name. |
| has_base: whether the extracted files have a base folder. This flag is used when checking if the existing |
| folder is a result of `extractall`, if it is, the extraction is skipped. For example, if A.zip is unzipped |
| to folder structure `A/*.png`, this flag should be True; if B.zip is unzipped to `*.png`, this flag should |
| be False. |
| progress: whether to display progress bar. |
| """ |
| with tempfile.TemporaryDirectory() as tmp_dir: |
| filename = filepath or Path(tmp_dir, _basename(url)).resolve() |
| download_url(url=url, filepath=filename, hash_val=hash_val, hash_type=hash_type, progress=progress) |
| extractall(filepath=filename, output_dir=output_dir, file_type=file_type, has_base=has_base) |
|
|