| """ |
| User script for AUTOMATIC111's SD WebUI that integrates MagicPrompt |
| Copyright (C) 2022 Spaceginner |
| |
| This user script is free software: you can redistribute it and/or modify |
| it under the terms of the GNU General Public License as published by |
| the Free Software Foundation, either version 3 of the License, or |
| (at your option) any later version. |
| |
| This user script is distributed in the hope that it will be useful, |
| but WITHOUT ANY WARRANTY; without even the implied warranty of |
| MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| GNU General Public License for more details. |
| |
| You should have received a copy of the GNU General Public License |
| along with this user script. If not, see <https://www.gnu.org/licenses/>. |
| |
| Contact me via Discord (Spaceginner#7688), email (ivan.demian2009@gmail.com) |
| or via "Issues" tab on Github page of this script |
| (https://github.com/Spaceginner/MagicPrompt-awebui) |
| |
| Credits: |
| Special thanks to Gustavosta for creating MagicPrompt AI model. |
| Also credits go to u/Letharguss (for creating basic script) |
| and SoCalGuitarist#2586 (for figuring out how to change prompt batch to batch) |
| |
| Version: 1.0.0 |
| """ |
|
|
| import os |
| import sys |
| import subprocess |
|
|
| |
| try: |
| from aitextgen import aitextgen |
| except: |
| print("[MagicPrompt script] aitextgen module is not found, downloading...") |
| if os.path.exists("./venv"): |
| subprocess.call(["./venv/Scripts/python", "-m", "pip", "-q", "--disable-pip-version-check", "--no-input", "install", "aitextgen"]) |
| else: |
| subprocess.call(["python", "-m", "pip", "-q", "--disable-pip-version-check", "--no-input", "install", "aitextgen"]) |
| print("[MagicPrompt script] aitextgen module is downloaded") |
|
|
| import gradio as gr |
| import torch |
|
|
| import modules.scripts as scripts |
| from modules.processing import Processed, process_images |
| from modules.shared import state |
|
|
| |
| if (not os.path.isdir("./models/MagicPrompt/")): |
| print("[MagicPrompt script] MagicPrompt model is not found, downloading MagicPrompt model...") |
| os.mkdir("./models/MagicPrompt/") |
| subprocess.call(["git", "clone", "--quiet", "https://huggingface.co/Gustavosta/MagicPrompt-Stable-Diffusion", "./models/MagicPrompt/."]) |
| print("[MagicPrompt script] MagicPrompt model is downloaded") |
|
|
| def getOrdinalNum(n): |
| if str(n)[-1] == "1": |
| return f"{n}st" |
| elif str(n)[-1] == "2": |
| return f"{n}nd" |
| elif str(n)[-1] == "3": |
| return f"{n}rd" |
| else: |
| return f"{n}th" |
|
|
| class Script(scripts.Script): |
| |
| gpt = None |
|
|
| def title(self): |
| return "MagicPrompt" |
|
|
| def show(self, isImg2img): |
| |
| return not isImg2img |
|
|
| def ui(self, isImg2img): |
| |
| promptLength = gr.Slider(label="Prompt max. length", value=75, minimum=1, maximum=300, step=1) |
| temp = gr.Slider(label="Temperature", value=0.7, minimum=0.1, maximum=2, step=0.1) |
| useSameSeed = gr.Checkbox(label="Use same seed for each batch", value=False) |
| useUniquePrompt = gr.Checkbox(label="Use unique prompt for each batch", value=True) |
| isPrioritized = gr.Checkbox(label="Iniatial prompt will have more prority over generated one", value=False) |
| doPregenerating = gr.Checkbox(label="Enable prompt pregenerating (Theoretical perfomance boost). If you dont know how many images do you want to generate, disable it", value=True) |
| doUnloadModel = gr.Checkbox(label="Unload MagicPrompt model from VRAM/RAM after this run. (Decreased perfomance between runs, as it need to load again)", value=False) |
| |
| return [promptLength, temp, useSameSeed, useUniquePrompt, isPrioritized, doPregenerating, doUnloadModel] |
|
|
| def run(self, p, promptLength, temp, useSameSeed, useUniquePrompt, isPrioritized, doPregenerating, doUnloadModel): |
| print() |
|
|
| |
| if type(self.gpt) != aitextgen: |
| self.gpt = aitextgen(model_folder="./models/MagicPrompt/", tokenizer_file="./models/MagicPrompt/tokenizer.json", to_gpu=torch.cuda.is_available()) |
|
|
| p.do_not_save_grid = True |
|
|
| |
| p.prompt = p.prompt[0] if type(p.prompt) == list else p.prompt |
|
|
| |
| |
| state.job_count = p.n_iter |
| p.n_iter = 1 |
|
|
| |
| originalPrompt = p.prompt |
| if (originalPrompt != "" and isPrioritized): |
| originalPrompt = "(" + originalPrompt + ")" |
|
|
| |
| prompts = [] |
| if (doPregenerating): |
| print(f"[MagicPrompt script] Pregenerating prompt{'s' if state.job_count > 1 else ''}...") |
| for i in range(state.job_count): |
| if (i == 0 or useUniquePrompt): |
| if state.interrupted: |
| print(f"[MagicPrompt script] Pregeneration interrupted") |
| break |
| |
| prompts.append(self.gpt.generate_one(prompt=originalPrompt, max_length=promptLength, temperature=temp)) |
| if state.job_count > 1: |
| print(f"[MagicPrompt script] Pregenerated {getOrdinalNum(i+1)} prompt...") |
| else: |
| break |
| print("[MagicPrompt script] Pregenerating finished") |
|
|
| images = [] |
| for i in range(state.job_count): |
| if state.skipped: |
| print("Rendering of current batch skipped") |
| continue |
|
|
| if state.interrupted: |
| print(f"Rendering interrupted") |
| break |
| |
| state.job = f"{i+1} out of {state.job_count}" |
|
|
| |
| sys.stdout.write('\033[2K\033[1G') |
| print("\n") |
|
|
| |
| if (i == 0 or useUniquePrompt): |
| if doPregenerating: |
| p.prompt = prompts[i] |
| else: |
| |
| print(f"[MagicPrompt script] Generating prompt for {getOrdinalNum(i+1)} batch...") |
| p.prompt = self.gpt.generate_one(prompt=originalPrompt, max_length=promptLength, temperature=temp) |
|
|
| print(f"[MagicPrompt script] Generated prompt for {getOrdinalNum(i+1)} batch: {p.prompt}") |
|
|
| |
| images += process_images(p).images |
|
|
| if not useSameSeed: |
| if not p.seed == -1: |
| p.seed += 1 |
|
|
| |
| if doUnloadModel: |
| del self.gpt |
|
|
| return Processed(p, images, p.seed, "") |
|
|