| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| import copy |
| import os |
| import pickle |
| import tempfile |
| from types import ModuleType |
| from typing import Any, Hashable |
|
|
| import torch |
| from torch.serialization import DEFAULT_PROTOCOL |
|
|
| from monai.config.type_definitions import PathLike |
|
|
| __all__ = ["StateCacher"] |
|
|
|
|
| class StateCacher: |
| """Class to cache and retrieve the state of an object. |
| |
| Objects can either be stored in memory or on disk. If stored on disk, they can be |
| stored in a given directory, or alternatively a temporary location will be used. |
| |
| If necessary/possible, restored objects will be returned to their original device. |
| |
| Example: |
| |
| >>> state_cacher = StateCacher(memory_cache, cache_dir=cache_dir) |
| >>> state_cacher.store("model", model.state_dict()) |
| >>> model.load_state_dict(state_cacher.retrieve("model")) |
| """ |
|
|
| def __init__( |
| self, |
| in_memory: bool, |
| cache_dir: PathLike | None = None, |
| allow_overwrite: bool = True, |
| pickle_module: ModuleType = pickle, |
| pickle_protocol: int = DEFAULT_PROTOCOL, |
| ) -> None: |
| """Constructor. |
| |
| Args: |
| in_memory: boolean to determine if the object will be cached in memory or on |
| disk. |
| cache_dir: directory for data to be cached if `in_memory==False`. Defaults |
| to using a temporary directory. Any created files will be deleted during |
| the `StateCacher`'s destructor. |
| allow_overwrite: allow the cache to be overwritten. If set to `False`, an |
| error will be thrown if a matching already exists in the list of cached |
| objects. |
| pickle_module: module used for pickling metadata and objects, default to `pickle`. |
| this arg is used by `torch.save`, for more details, please check: |
| https://pytorch.org/docs/stable/generated/torch.save.html#torch.save. |
| pickle_protocol: can be specified to override the default protocol, default to `2`. |
| this arg is used by `torch.save`, for more details, please check: |
| https://pytorch.org/docs/stable/generated/torch.save.html#torch.save. |
| |
| """ |
| self.in_memory = in_memory |
| self.cache_dir = tempfile.gettempdir() if cache_dir is None else cache_dir |
| if not os.path.isdir(self.cache_dir): |
| raise ValueError("Given `cache_dir` is not a valid directory.") |
|
|
| self.allow_overwrite = allow_overwrite |
| self.pickle_module = pickle_module |
| self.pickle_protocol = pickle_protocol |
| self.cached: dict = {} |
|
|
| def store( |
| self, key: Hashable, data_obj: Any, pickle_module: ModuleType | None = None, pickle_protocol: int | None = None |
| ) -> None: |
| """ |
| Store a given object with the given key name. |
| |
| Args: |
| key: key of the data object to store. |
| data_obj: data object to store. |
| pickle_module: module used for pickling metadata and objects, default to `self.pickle_module`. |
| this arg is used by `torch.save`, for more details, please check: |
| https://pytorch.org/docs/stable/generated/torch.save.html#torch.save. |
| pickle_protocol: can be specified to override the default protocol, default to `self.pickle_protocol`. |
| this arg is used by `torch.save`, for more details, please check: |
| https://pytorch.org/docs/stable/generated/torch.save.html#torch.save. |
| |
| """ |
| if key in self.cached and not self.allow_overwrite: |
| raise RuntimeError("Cached key already exists and overwriting is disabled.") |
| if self.in_memory: |
| self.cached.update({key: {"obj": copy.deepcopy(data_obj)}}) |
| else: |
| fn = os.path.join(self.cache_dir, f"state_{key}_{id(self)}.pt") |
| self.cached.update({key: {"obj": fn}}) |
| torch.save( |
| obj=data_obj, |
| f=fn, |
| pickle_module=self.pickle_module if pickle_module is None else pickle_module, |
| pickle_protocol=self.pickle_protocol if pickle_protocol is None else pickle_protocol, |
| ) |
| |
| if hasattr(data_obj, "device"): |
| self.cached[key]["device"] = data_obj.device |
|
|
| def retrieve(self, key: Hashable) -> Any: |
| """Retrieve the object stored under a given key name.""" |
| if key not in self.cached: |
| raise KeyError(f"Target {key} was not cached.") |
|
|
| if self.in_memory: |
| return self.cached[key]["obj"] |
|
|
| fn = self.cached[key]["obj"] |
| if not os.path.exists(fn): |
| raise RuntimeError(f"Failed to load state in {fn}. File doesn't exist anymore.") |
| data_obj = torch.load(fn, map_location=lambda storage, location: storage) |
| |
| if "device" in self.cached[key]: |
| data_obj = data_obj.to(self.cached[key]["device"]) |
| return data_obj |
|
|
| def __del__(self): |
| """If necessary, delete any cached files existing in `cache_dir`.""" |
| if not self.in_memory: |
| for k in self.cached: |
| if os.path.exists(self.cached[k]["obj"]): |
| os.remove(self.cached[k]["obj"]) |
|
|