| from __future__ import annotations |
|
|
| import configparser |
| import os |
| import threading |
| import re |
|
|
| from modules import shared, errors, cache, scripts |
| from modules.gitpython_hack import Repo |
| from modules.paths_internal import extensions_dir, extensions_builtin_dir, script_path |
|
|
|
|
| os.makedirs(extensions_dir, exist_ok=True) |
|
|
|
|
| def active(): |
| if shared.cmd_opts.disable_all_extensions or shared.opts.disable_all_extensions == "all": |
| return [] |
| elif shared.cmd_opts.disable_extra_extensions or shared.opts.disable_all_extensions == "extra": |
| return [x for x in extensions if x.enabled and x.is_builtin] |
| else: |
| return [x for x in extensions if x.enabled] |
|
|
|
|
| class ExtensionMetadata: |
| filename = "metadata.ini" |
| config: configparser.ConfigParser |
| canonical_name: str |
| requires: list |
|
|
| def __init__(self, path, canonical_name): |
| self.config = configparser.ConfigParser() |
|
|
| filepath = os.path.join(path, self.filename) |
| if os.path.isfile(filepath): |
| try: |
| self.config.read(filepath) |
| except Exception: |
| errors.report(f"Error reading {self.filename} for extension {canonical_name}.", exc_info=True) |
|
|
| self.canonical_name = self.config.get("Extension", "Name", fallback=canonical_name) |
| self.canonical_name = canonical_name.lower().strip() |
|
|
| self.requires = self.get_script_requirements("Requires", "Extension") |
|
|
| def get_script_requirements(self, field, section, extra_section=None): |
| """reads a list of requirements from the config; field is the name of the field in the ini file, |
| like Requires or Before, and section is the name of the [section] in the ini file; additionally, |
| reads more requirements from [extra_section] if specified.""" |
|
|
| x = self.config.get(section, field, fallback='') |
|
|
| if extra_section: |
| x = x + ', ' + self.config.get(extra_section, field, fallback='') |
|
|
| return self.parse_list(x.lower()) |
|
|
| def parse_list(self, text): |
| """converts a line from config ("ext1 ext2, ext3 ") into a python list (["ext1", "ext2", "ext3"])""" |
|
|
| if not text: |
| return [] |
|
|
| |
| return [x for x in re.split(r"[,\s]+", text.strip()) if x] |
|
|
|
|
| class Extension: |
| lock = threading.Lock() |
| cached_fields = ['remote', 'commit_date', 'branch', 'commit_hash', 'version'] |
| metadata: ExtensionMetadata |
|
|
| def __init__(self, name, path, enabled=True, is_builtin=False, metadata=None): |
| self.name = name |
| self.path = path |
| self.enabled = enabled |
| self.status = '' |
| self.can_update = False |
| self.is_builtin = is_builtin |
| self.commit_hash = '' |
| self.commit_date = None |
| self.version = '' |
| self.branch = None |
| self.remote = None |
| self.have_info_from_repo = False |
| self.metadata = metadata if metadata else ExtensionMetadata(self.path, name.lower()) |
| self.canonical_name = metadata.canonical_name |
|
|
| def to_dict(self): |
| return {x: getattr(self, x) for x in self.cached_fields} |
|
|
| def from_dict(self, d): |
| for field in self.cached_fields: |
| setattr(self, field, d[field]) |
|
|
| def read_info_from_repo(self): |
| if self.is_builtin or self.have_info_from_repo: |
| return |
|
|
| def read_from_repo(): |
| with self.lock: |
| if self.have_info_from_repo: |
| return |
|
|
| self.do_read_info_from_repo() |
|
|
| return self.to_dict() |
|
|
| try: |
| d = cache.cached_data_for_file('extensions-git', self.name, os.path.join(self.path, ".git"), read_from_repo) |
| self.from_dict(d) |
| except FileNotFoundError: |
| pass |
| self.status = 'unknown' if self.status == '' else self.status |
|
|
| def do_read_info_from_repo(self): |
| repo = None |
| try: |
| if os.path.exists(os.path.join(self.path, ".git")): |
| repo = Repo(self.path) |
| except Exception: |
| errors.report(f"Error reading github repository info from {self.path}", exc_info=True) |
|
|
| if repo is None or repo.bare: |
| self.remote = None |
| else: |
| try: |
| self.remote = next(repo.remote().urls, None) |
| commit = repo.head.commit |
| self.commit_date = commit.committed_date |
| if repo.active_branch: |
| self.branch = repo.active_branch.name |
| self.commit_hash = commit.hexsha |
| self.version = self.commit_hash[:8] |
|
|
| except Exception: |
| errors.report(f"Failed reading extension data from Git repository ({self.name})", exc_info=True) |
| self.remote = None |
|
|
| self.have_info_from_repo = True |
|
|
| def list_files(self, subdir, extension): |
| dirpath = os.path.join(self.path, subdir) |
| if not os.path.isdir(dirpath): |
| return [] |
|
|
| res = [] |
| for filename in sorted(os.listdir(dirpath)): |
| res.append(scripts.ScriptFile(self.path, filename, os.path.join(dirpath, filename))) |
|
|
| res = [x for x in res if os.path.splitext(x.path)[1].lower() == extension and os.path.isfile(x.path)] |
|
|
| return res |
|
|
| def check_updates(self): |
| repo = Repo(self.path) |
| for fetch in repo.remote().fetch(dry_run=True): |
| if fetch.flags != fetch.HEAD_UPTODATE: |
| self.can_update = True |
| self.status = "new commits" |
| return |
|
|
| try: |
| origin = repo.rev_parse('origin') |
| if repo.head.commit != origin: |
| self.can_update = True |
| self.status = "behind HEAD" |
| return |
| except Exception: |
| self.can_update = False |
| self.status = "unknown (remote error)" |
| return |
|
|
| self.can_update = False |
| self.status = "latest" |
|
|
| def fetch_and_reset_hard(self, commit='origin'): |
| repo = Repo(self.path) |
| |
| |
| repo.git.fetch(all=True) |
| repo.git.reset(commit, hard=True) |
| self.have_info_from_repo = False |
|
|
|
|
| def list_extensions(): |
| extensions.clear() |
|
|
| if shared.cmd_opts.disable_all_extensions: |
| print("*** \"--disable-all-extensions\" arg was used, will not load any extensions ***") |
| elif shared.opts.disable_all_extensions == "all": |
| print("*** \"Disable all extensions\" option was set, will not load any extensions ***") |
| elif shared.cmd_opts.disable_extra_extensions: |
| print("*** \"--disable-extra-extensions\" arg was used, will only load built-in extensions ***") |
| elif shared.opts.disable_all_extensions == "extra": |
| print("*** \"Disable all extensions\" option was set, will only load built-in extensions ***") |
|
|
| loaded_extensions = {} |
|
|
| |
| for dirname in [extensions_builtin_dir, extensions_dir]: |
| if not os.path.isdir(dirname): |
| continue |
|
|
| for extension_dirname in sorted(os.listdir(dirname)): |
| path = os.path.join(dirname, extension_dirname) |
| if not os.path.isdir(path): |
| continue |
|
|
| canonical_name = extension_dirname |
| metadata = ExtensionMetadata(path, canonical_name) |
|
|
| |
| already_loaded_extension = loaded_extensions.get(metadata.canonical_name) |
| if already_loaded_extension is not None: |
| errors.report(f'Duplicate canonical name "{canonical_name}" found in extensions "{extension_dirname}" and "{already_loaded_extension.name}". Former will be discarded.', exc_info=False) |
| continue |
|
|
| is_builtin = dirname == extensions_builtin_dir |
| extension = Extension(name=extension_dirname, path=path, enabled=extension_dirname not in shared.opts.disabled_extensions, is_builtin=is_builtin, metadata=metadata) |
| extensions.append(extension) |
| loaded_extensions[canonical_name] = extension |
|
|
| |
| for extension in extensions: |
| for req in extension.metadata.requires: |
| required_extension = loaded_extensions.get(req) |
| if required_extension is None: |
| errors.report(f'Extension "{extension.name}" requires "{req}" which is not installed.', exc_info=False) |
| continue |
|
|
| if not extension.enabled: |
| errors.report(f'Extension "{extension.name}" requires "{required_extension.name}" which is disabled.', exc_info=False) |
| continue |
|
|
|
|
| extensions: list[Extension] = [] |
|
|