Datasets:
Tags:
Not-For-All-Audiences
Upload anima_lora_train.ipynb
Browse files- anima_lora_train.ipynb +1 -0
anima_lora_train.ipynb
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"cells":[{"cell_type":"code","execution_count":1,"metadata":{"colab":{"base_uri":"https://localhost:8080/"},"executionInfo":{"elapsed":13098,"status":"ok","timestamp":1775914578871,"user":{"displayName":"No Name","userId":"10578412414437288386"},"user_tz":-120},"id":"xCouXHPohJxN","outputId":"9f8509bb-4b0e-4238-ba9a-e518521eb0eb"},"outputs":[{"output_type":"stream","name":"stdout","text":["Mounted at /content/drive\n","โ
HF_TOKEN loaded from Colab secrets\n","๐ Ready for downloads and training!\n"]}],"source":["#@markdown # Cell 1: Mount Google Drive + Load HF Token\n","#@markdown Mounts your Drive and securely loads your Hugging Face token (Colab Secrets preferred).\n","\n","from google.colab import drive\n","drive.mount('/content/drive')\n","\n","# Load HF_TOKEN (Colab secrets preferred, fallback to input)\n","try:\n"," from google.colab import userdata\n"," hf_token = userdata.get('HF_TOKEN')\n"," print(\"โ
HF_TOKEN loaded from Colab secrets\")\n","except:\n"," import getpass\n"," hf_token = getpass.getpass(\"Enter your Hugging Face Token (or leave blank): \")\n"," if hf_token.strip() == \"\":\n"," hf_token = None\n"," print(\"โ
HF_TOKEN set\")\n","\n","import os\n","os.environ[\"HF_TOKEN\"] = hf_token or \"\"\n","print(\"๐ Ready for downloads and training!\")"]},{"cell_type":"code","execution_count":2,"metadata":{"colab":{"base_uri":"https://localhost:8080/"},"executionInfo":{"elapsed":16305,"status":"ok","timestamp":1775914595179,"user":{"displayName":"No Name","userId":"10578412414437288386"},"user_tz":-120},"id":"W6zIOTKihLRy","outputId":"4cf4a21e-259c-4a3a-883c-b0202311fb0d"},"outputs":[{"output_type":"stream","name":"stdout","text":["\r0% [Working]\r \rGet:1 https://cli.github.com/packages stable InRelease [3,917 B]\n","\r0% [Connecting to archive.ubuntu.com (185.125.190.83)] [Connecting to security.\r0% [Connecting to archive.ubuntu.com (185.125.190.83)] [Connecting to security.\r \rGet:2 https://cloud.r-project.org/bin/linux/ubuntu jammy-cran40/ InRelease [3,632 B]\n","Get:3 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 InRelease [1,581 B]\n","Get:4 https://cloud.r-project.org/bin/linux/ubuntu jammy-cran40/ Packages [88.5 kB]\n","Get:5 https://r2u.stat.illinois.edu/ubuntu jammy InRelease [6,555 B]\n","Hit:6 http://archive.ubuntu.com/ubuntu jammy InRelease\n","Get:7 http://security.ubuntu.com/ubuntu jammy-security InRelease [129 kB]\n","Get:8 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages [2,497 kB]\n","Get:9 http://archive.ubuntu.com/ubuntu jammy-updates InRelease [128 kB]\n","Get:10 https://r2u.stat.illinois.edu/ubuntu jammy/main amd64 Packages [2,965 kB]\n","Get:11 https://ppa.launchpadcontent.net/deadsnakes/ppa/ubuntu jammy InRelease [18.1 kB]\n","Hit:12 https://ppa.launchpadcontent.net/graphics-drivers/ppa/ubuntu jammy InRelease\n","Get:13 https://r2u.stat.illinois.edu/ubuntu jammy/main all Packages [10.0 MB]\n","Get:14 http://security.ubuntu.com/ubuntu jammy-security/restricted amd64 Packages [6,917 kB]\n","Get:15 http://archive.ubuntu.com/ubuntu jammy-backports InRelease [127 kB]\n","Hit:16 https://ppa.launchpadcontent.net/ubuntugis/ppa/ubuntu jammy InRelease\n","Get:17 https://ppa.launchpadcontent.net/deadsnakes/ppa/ubuntu jammy/main amd64 Packages [38.8 kB]\n","Get:18 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 Packages [4,226 kB]\n","Get:19 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 Packages [1,622 kB]\n","Get:20 http://security.ubuntu.com/ubuntu jammy-security/main amd64 Packages [3,889 kB]\n","Get:21 http://archive.ubuntu.com/ubuntu jammy-updates/restricted amd64 Packages [7,143 kB]\n","Get:22 http://security.ubuntu.com/ubuntu jammy-security/universe amd64 Packages [1,311 kB]\n","Fetched 41.1 MB in 3s (12.5 MB/s)\n","Reading package lists... Done\n","W: Skipping acquire of configured file 'main/source/Sources' as repository 'https://r2u.stat.illinois.edu/ubuntu jammy InRelease' does not seem to provide it (sources.list entry misspelt?)\n","Reading package lists... Done\n","Building dependency tree... Done\n","Reading state information... Done\n","git is already the newest version (1:2.34.1-1ubuntu1.17).\n","The following additional packages will be installed:\n"," libaria2-0 libc-ares2 python3-pip-whl python3-setuptools-whl\n","The following NEW packages will be installed:\n"," aria2 libaria2-0 libc-ares2 python3-pip-whl python3-setuptools-whl\n"," python3.10-venv\n","0 upgraded, 6 newly installed, 0 to remove and 74 not upgraded.\n","Need to get 3,994 kB of archives.\n","After this operation, 8,194 kB of additional disk space will be used.\n","Get:1 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libc-ares2 amd64 1.18.1-1ubuntu0.22.04.3 [45.1 kB]\n","Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libaria2-0 amd64 1.36.0-1 [1,086 kB]\n","Get:3 https://ppa.launchpadcontent.net/ubuntugis/ppa/ubuntu jammy/main amd64 python3-setuptools-whl all 68.1.2-2~jammy3 [792 kB]\n","Get:4 http://archive.ubuntu.com/ubuntu jammy/universe amd64 aria2 amd64 1.36.0-1 [381 kB]\n","Get:5 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-pip-whl all 22.0.2+dfsg-1ubuntu0.7 [1,683 kB]\n","Get:6 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3.10-venv amd64 3.10.12-1~22.04.15 [5,714 B]\n","Fetched 3,994 kB in 3s (1,549 kB/s)\n","Selecting previously unselected package libc-ares2:amd64.\n","(Reading database ... 122354 files and directories currently installed.)\n","Preparing to unpack .../0-libc-ares2_1.18.1-1ubuntu0.22.04.3_amd64.deb ...\n","Unpacking libc-ares2:amd64 (1.18.1-1ubuntu0.22.04.3) ...\n","Selecting previously unselected package libaria2-0:amd64.\n","Preparing to unpack .../1-libaria2-0_1.36.0-1_amd64.deb ...\n","Unpacking libaria2-0:amd64 (1.36.0-1) ...\n","Selecting previously unselected package aria2.\n","Preparing to unpack .../2-aria2_1.36.0-1_amd64.deb ...\n","Unpacking aria2 (1.36.0-1) ...\n","Selecting previously unselected package python3-pip-whl.\n","Preparing to unpack .../3-python3-pip-whl_22.0.2+dfsg-1ubuntu0.7_all.deb ...\n","Unpacking python3-pip-whl (22.0.2+dfsg-1ubuntu0.7) ...\n","Selecting previously unselected package python3-setuptools-whl.\n","Preparing to unpack .../4-python3-setuptools-whl_68.1.2-2~jammy3_all.deb ...\n","Unpacking python3-setuptools-whl (68.1.2-2~jammy3) ...\n","Selecting previously unselected package python3.10-venv.\n","Preparing to unpack .../5-python3.10-venv_3.10.12-1~22.04.15_amd64.deb ...\n","Unpacking python3.10-venv (3.10.12-1~22.04.15) ...\n","Setting up python3-setuptools-whl (68.1.2-2~jammy3) ...\n","Setting up python3-pip-whl (22.0.2+dfsg-1ubuntu0.7) ...\n","Setting up libc-ares2:amd64 (1.18.1-1ubuntu0.22.04.3) ...\n","Setting up libaria2-0:amd64 (1.36.0-1) ...\n","Setting up python3.10-venv (3.10.12-1~22.04.15) ...\n","Setting up aria2 (1.36.0-1) ...\n","Processing triggers for man-db (2.10.2-1) ...\n","Processing triggers for libc-bin (2.35-0ubuntu3.8) ...\n","/sbin/ldconfig.real: /usr/local/lib/libtbbmalloc_proxy.so.2 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libumf.so.1 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libur_loader.so.0 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libtcm.so.1 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libtcm_debug.so.1 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libtbbmalloc.so.2 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libur_adapter_level_zero_v2.so.0 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libtbbbind_2_5.so.3 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libtbbbind_2_0.so.3 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libur_adapter_level_zero.so.0 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libur_adapter_opencl.so.0 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libtbbbind.so.3 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libtbb.so.12 is not a symbolic link\n","\n","/sbin/ldconfig.real: /usr/local/lib/libhwloc.so.15 is not a symbolic link\n","\n"]}],"source":["#@markdown # Cell 2: Install System Packages\n","#@markdown Updates apt and installs aria2 + git (required for fast downloads and cloning).\n","\n","!apt-get update -y && apt-get install -y aria2 git python3.10-venv"]},{"cell_type":"code","execution_count":3,"metadata":{"cellView":"form","colab":{"base_uri":"https://localhost:8080/"},"executionInfo":{"elapsed":250301,"status":"ok","timestamp":1775914845481,"user":{"displayName":"No Name","userId":"10578412414437288386"},"user_tz":-120},"id":"tvLBG0b4jV1J","outputId":"e8510b6b-c0b1-47d4-84e5-6bde18093c72"},"outputs":[{"output_type":"stream","name":"stdout","text":["/content\n","Cloning into 'sd-scripts'...\n","remote: Enumerating objects: 11050, done.\u001b[K\n","remote: Counting objects: 100% (146/146), done.\u001b[K\n","remote: Compressing objects: 100% (107/107), done.\u001b[K\n","remote: Total 11050 (delta 91), reused 39 (delta 39), pack-reused 10904 (from 4)\u001b[K\n","Receiving objects: 100% (11050/11050), 17.56 MiB | 23.11 MiB/s, done.\n","Resolving deltas: 100% (7862/7862), done.\n","/content/sd-scripts\n","Looking in indexes: https://download.pytorch.org/whl/cu121\n","Collecting torch==2.5.1+cu121\n"," Downloading https://download-r2.pytorch.org/whl/cu121/torch-2.5.1%2Bcu121-cp312-cp312-linux_x86_64.whl (780.4 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m780.4/780.4 MB\u001b[0m \u001b[31m2.0 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting torchvision==0.20.1+cu121\n"," Downloading https://download-r2.pytorch.org/whl/cu121/torchvision-0.20.1%2Bcu121-cp312-cp312-linux_x86_64.whl (7.3 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโ๏ฟฝ๏ฟฝโโโโโโโโโโโ\u001b[0m \u001b[32m7.3/7.3 MB\u001b[0m \u001b[31m88.1 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting torchaudio==2.5.1+cu121\n"," Downloading https://download-r2.pytorch.org/whl/cu121/torchaudio-2.5.1%2Bcu121-cp312-cp312-linux_x86_64.whl (3.4 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m3.4/3.4 MB\u001b[0m \u001b[31m88.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hRequirement already satisfied: filelock in /usr/local/lib/python3.12/dist-packages (from torch==2.5.1+cu121) (3.25.2)\n","Requirement already satisfied: typing-extensions>=4.8.0 in /usr/local/lib/python3.12/dist-packages (from torch==2.5.1+cu121) (4.15.0)\n","Requirement already satisfied: networkx in /usr/local/lib/python3.12/dist-packages (from torch==2.5.1+cu121) (3.6.1)\n","Requirement already satisfied: jinja2 in /usr/local/lib/python3.12/dist-packages (from torch==2.5.1+cu121) (3.1.6)\n","Requirement already satisfied: fsspec in /usr/local/lib/python3.12/dist-packages (from torch==2.5.1+cu121) (2025.3.0)\n","Collecting nvidia-cuda-nvrtc-cu12==12.1.105 (from torch==2.5.1+cu121)\n"," Downloading https://download.pytorch.org/whl/cu121/nvidia_cuda_nvrtc_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (23.7 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m23.7/23.7 MB\u001b[0m \u001b[31m71.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting nvidia-cuda-runtime-cu12==12.1.105 (from torch==2.5.1+cu121)\n"," Downloading https://download.pytorch.org/whl/cu121/nvidia_cuda_runtime_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (823 kB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m823.6/823.6 kB\u001b[0m \u001b[31m38.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting nvidia-cuda-cupti-cu12==12.1.105 (from torch==2.5.1+cu121)\n"," Downloading https://download.pytorch.org/whl/cu121/nvidia_cuda_cupti_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (14.1 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m14.1/14.1 MB\u001b[0m \u001b[31m89.2 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting nvidia-cudnn-cu12==9.1.0.70 (from torch==2.5.1+cu121)\n"," Downloading https://download.pytorch.org/whl/cu121/nvidia_cudnn_cu12-9.1.0.70-py3-none-manylinux2014_x86_64.whl (664.8 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m664.8/664.8 MB\u001b[0m \u001b[31m882.0 kB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting nvidia-cublas-cu12==12.1.3.1 (from torch==2.5.1+cu121)\n"," Downloading https://download.pytorch.org/whl/cu121/nvidia_cublas_cu12-12.1.3.1-py3-none-manylinux1_x86_64.whl (410.6 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m410.6/410.6 MB\u001b[0m \u001b[31m1.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting nvidia-cufft-cu12==11.0.2.54 (from torch==2.5.1+cu121)\n"," Downloading https://download.pytorch.org/whl/cu121/nvidia_cufft_cu12-11.0.2.54-py3-none-manylinux1_x86_64.whl (121.6 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m121.6/121.6 MB\u001b[0m \u001b[31m7.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting nvidia-curand-cu12==10.3.2.106 (from torch==2.5.1+cu121)\n"," Downloading https://download.pytorch.org/whl/cu121/nvidia_curand_cu12-10.3.2.106-py3-none-manylinux1_x86_64.whl (56.5 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m56.5/56.5 MB\u001b[0m \u001b[31m14.2 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting nvidia-cusolver-cu12==11.4.5.107 (from torch==2.5.1+cu121)\n"," Downloading https://download.pytorch.org/whl/cu121/nvidia_cusolver_cu12-11.4.5.107-py3-none-manylinux1_x86_64.whl (124.2 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m124.2/124.2 MB\u001b[0m \u001b[31m7.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting nvidia-cusparse-cu12==12.1.0.106 (from torch==2.5.1+cu121)\n"," Downloading https://download.pytorch.org/whl/cu121/nvidia_cusparse_cu12-12.1.0.106-py3-none-manylinux1_x86_64.whl (196.0 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m196.0/196.0 MB\u001b[0m \u001b[31m6.1 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting nvidia-nccl-cu12==2.21.5 (from torch==2.5.1+cu121)\n"," Downloading https://download.pytorch.org/whl/nvidia_nccl_cu12-2.21.5-py3-none-manylinux2014_x86_64.whl (188.7 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m188.7/188.7 MB\u001b[0m \u001b[31m6.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting nvidia-nvtx-cu12==12.1.105 (from torch==2.5.1+cu121)\n"," Downloading https://download.pytorch.org/whl/cu121/nvidia_nvtx_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (99 kB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m99.1/99.1 kB\u001b[0m \u001b[31m12.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hCollecting triton==3.1.0 (from torch==2.5.1+cu121)\n"," Downloading https://download-r2.pytorch.org/whl/triton-3.1.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (209.6 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m209.6/209.6 MB\u001b[0m \u001b[31m5.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hRequirement already satisfied: setuptools in /usr/local/lib/python3.12/dist-packages (from torch==2.5.1+cu121) (75.2.0)\n","Collecting sympy==1.13.1 (from torch==2.5.1+cu121)\n"," Downloading sympy-1.13.1-py3-none-any.whl.metadata (12 kB)\n","Requirement already satisfied: numpy in /usr/local/lib/python3.12/dist-packages (from torchvision==0.20.1+cu121) (2.0.2)\n","Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /usr/local/lib/python3.12/dist-packages (from torchvision==0.20.1+cu121) (11.3.0)\n","Requirement already satisfied: nvidia-nvjitlink-cu12 in /usr/local/lib/python3.12/dist-packages (from nvidia-cusolver-cu12==11.4.5.107->torch==2.5.1+cu121) (12.8.93)\n","Requirement already satisfied: mpmath<1.4,>=1.1.0 in /usr/local/lib/python3.12/dist-packages (from sympy==1.13.1->torch==2.5.1+cu121) (1.3.0)\n","Requirement already satisfied: MarkupSafe>=2.0 in /usr/local/lib/python3.12/dist-packages (from jinja2->torch==2.5.1+cu121) (3.0.3)\n","Downloading sympy-1.13.1-py3-none-any.whl (6.2 MB)\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m6.2/6.2 MB\u001b[0m \u001b[31m60.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25hInstalling collected packages: triton, sympy, nvidia-nvtx-cu12, nvidia-nccl-cu12, nvidia-cusparse-cu12, nvidia-curand-cu12, nvidia-cufft-cu12, nvidia-cuda-runtime-cu12, nvidia-cuda-nvrtc-cu12, nvidia-cuda-cupti-cu12, nvidia-cublas-cu12, nvidia-cusolver-cu12, nvidia-cudnn-cu12, torch, torchvision, torchaudio\n"," Attempting uninstall: triton\n"," Found existing installation: triton 3.6.0\n"," Uninstalling triton-3.6.0:\n"," Successfully uninstalled triton-3.6.0\n"," Attempting uninstall: sympy\n"," Found existing installation: sympy 1.14.0\n"," Uninstalling sympy-1.14.0:\n"," Successfully uninstalled sympy-1.14.0\n"," Attempting uninstall: nvidia-nvtx-cu12\n"," Found existing installation: nvidia-nvtx-cu12 12.8.90\n"," Uninstalling nvidia-nvtx-cu12-12.8.90:\n"," Successfully uninstalled nvidia-nvtx-cu12-12.8.90\n"," Attempting uninstall: nvidia-nccl-cu12\n"," Found existing installation: nvidia-nccl-cu12 2.27.5\n"," Uninstalling nvidia-nccl-cu12-2.27.5:\n"," Successfully uninstalled nvidia-nccl-cu12-2.27.5\n"," Attempting uninstall: nvidia-cusparse-cu12\n"," Found existing installation: nvidia-cusparse-cu12 12.5.8.93\n"," Uninstalling nvidia-cusparse-cu12-12.5.8.93:\n"," Successfully uninstalled nvidia-cusparse-cu12-12.5.8.93\n"," Attempting uninstall: nvidia-curand-cu12\n"," Found existing installation: nvidia-curand-cu12 10.3.9.90\n"," Uninstalling nvidia-curand-cu12-10.3.9.90:\n"," Successfully uninstalled nvidia-curand-cu12-10.3.9.90\n"," Attempting uninstall: nvidia-cufft-cu12\n"," Found existing installation: nvidia-cufft-cu12 11.3.3.83\n"," Uninstalling nvidia-cufft-cu12-11.3.3.83:\n"," Successfully uninstalled nvidia-cufft-cu12-11.3.3.83\n"," Attempting uninstall: nvidia-cuda-runtime-cu12\n"," Found existing installation: nvidia-cuda-runtime-cu12 12.8.90\n"," Uninstalling nvidia-cuda-runtime-cu12-12.8.90:\n"," Successfully uninstalled nvidia-cuda-runtime-cu12-12.8.90\n"," Attempting uninstall: nvidia-cuda-nvrtc-cu12\n"," Found existing installation: nvidia-cuda-nvrtc-cu12 12.8.93\n"," Uninstalling nvidia-cuda-nvrtc-cu12-12.8.93:\n"," Successfully uninstalled nvidia-cuda-nvrtc-cu12-12.8.93\n"," Attempting uninstall: nvidia-cuda-cupti-cu12\n"," Found existing installation: nvidia-cuda-cupti-cu12 12.8.90\n"," Uninstalling nvidia-cuda-cupti-cu12-12.8.90:\n"," Successfully uninstalled nvidia-cuda-cupti-cu12-12.8.90\n"," Attempting uninstall: nvidia-cublas-cu12\n"," Found existing installation: nvidia-cublas-cu12 12.8.4.1\n"," Uninstalling nvidia-cublas-cu12-12.8.4.1:\n"," Successfully uninstalled nvidia-cublas-cu12-12.8.4.1\n"," Attempting uninstall: nvidia-cusolver-cu12\n"," Found existing installation: nvidia-cusolver-cu12 11.7.3.90\n"," Uninstalling nvidia-cusolver-cu12-11.7.3.90:\n"," Successfully uninstalled nvidia-cusolver-cu12-11.7.3.90\n"," Attempting uninstall: nvidia-cudnn-cu12\n"," Found existing installation: nvidia-cudnn-cu12 9.10.2.21\n"," Uninstalling nvidia-cudnn-cu12-9.10.2.21:\n"," Successfully uninstalled nvidia-cudnn-cu12-9.10.2.21\n"," Attempting uninstall: torch\n"," Found existing installation: torch 2.10.0+cu128\n"," Uninstalling torch-2.10.0+cu128:\n"," Successfully uninstalled torch-2.10.0+cu128\n"," Attempting uninstall: torchvision\n"," Found existing installation: torchvision 0.25.0+cu128\n"," Uninstalling torchvision-0.25.0+cu128:\n"," Successfully uninstalled torchvision-0.25.0+cu128\n"," Attempting uninstall: torchaudio\n"," Found existing installation: torchaudio 2.10.0+cu128\n"," Uninstalling torchaudio-2.10.0+cu128:\n"," Successfully uninstalled torchaudio-2.10.0+cu128\n","Successfully installed nvidia-cublas-cu12-12.1.3.1 nvidia-cuda-cupti-cu12-12.1.105 nvidia-cuda-nvrtc-cu12-12.1.105 nvidia-cuda-runtime-cu12-12.1.105 nvidia-cudnn-cu12-9.1.0.70 nvidia-cufft-cu12-11.0.2.54 nvidia-curand-cu12-10.3.2.106 nvidia-cusolver-cu12-11.4.5.107 nvidia-cusparse-cu12-12.1.0.106 nvidia-nccl-cu12-2.21.5 nvidia-nvtx-cu12-12.1.105 sympy-1.13.1 torch-2.5.1+cu121 torchaudio-2.5.1+cu121 torchvision-0.20.1+cu121 triton-3.1.0\n"," Preparing metadata (setup.py) ... \u001b[?25l\u001b[?25hdone\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m41.7/41.7 kB\u001b[0m \u001b[31m3.2 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25h Installing build dependencies ... \u001b[?25l\u001b[?25hdone\n"," Getting requirements to build wheel ... \u001b[?25l\u001b[?25hdone\n"," Preparing metadata (pyproject.toml) ... \u001b[?25l\u001b[?25hdone\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m71.0/71.0 kB\u001b[0m \u001b[31m6.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m354.7/354.7 kB\u001b[0m \u001b[31m23.2 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m11.2/11.2 MB\u001b[0m \u001b[31m117.1 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m3.2/3.2 MB\u001b[0m \u001b[31m111.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m44.8/44.8 kB\u001b[0m \u001b[31m4.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m62.5/62.5 MB\u001b[0m \u001b[31m11.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m44.6/44.6 kB\u001b[0m \u001b[31m4.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m287.4/287.4 kB\u001b[0m \u001b[31m29.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m434.8/434.8 kB\u001b[0m \u001b[31m38.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m558.8/558.8 kB\u001b[0m \u001b[31m41.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m243.4/243.4 kB\u001b[0m \u001b[31m25.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m60.7/60.7 MB\u001b[0m \u001b[31m18.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[2K \u001b[90mโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\u001b[0m \u001b[32m3.1/3.1 MB\u001b[0m \u001b[31m117.0 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n","\u001b[?25h Building wheel for schedulefree (pyproject.toml) ... \u001b[?25l\u001b[?25hdone\n","\u001b[31mERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.\n","bigframes 2.38.0 requires rich<14,>=12.4.4, but you have rich 14.1.0 which is incompatible.\u001b[0m\u001b[31m\n","\u001b[0mโ
sd-scripts + PyTorch 2.5.1 installed successfully โ anima_train_network.py is ready!\n","๐ก If you see any torch-related error later, restart the runtime (Runtime โ Restart session) and run this cell again.\n"]}],"source":["#@markdown # Cell 3: Clone kohya-ss/sd-scripts + Install PyTorch 2.5.1 (FIXED) + Requirements\n","#@markdown Clones the repo and installs the **compatible** PyTorch version for Colab + CUDA 12.1.\n","\n","%cd /content\n","!git clone https://github.com/kohya-ss/sd-scripts.git\n","%cd /content/sd-scripts\n","\n","# FIXED: Use 2.5.1+cu121 (the latest available on Colab's cu121 index and sufficient per kohya-ss docs)\n","!pip install torch==2.5.1+cu121 torchvision==0.20.1+cu121 torchaudio==2.5.1+cu121 --index-url https://download.pytorch.org/whl/cu121\n","\n","# Install the rest of the requirements\n","!pip install -q -r requirements.txt\n","\n","print(\"โ
sd-scripts + PyTorch 2.5.1 installed successfully โ anima_train_network.py is ready!\")\n","print(\"๐ก If you see any torch-related error later, restart the runtime (Runtime โ Restart session) and run this cell again.\")"]},{"cell_type":"code","execution_count":4,"metadata":{"colab":{"base_uri":"https://localhost:8080/"},"executionInfo":{"elapsed":929,"status":"ok","timestamp":1775914846414,"user":{"displayName":"No Name","userId":"10578412414437288386"},"user_tz":-120},"id":"mHd0DL_uhQpr","outputId":"2e9911ae-f095-4ff9-b023-80dd5fc9f094"},"outputs":[{"output_type":"stream","name":"stdout","text":["๐ฆ Unzipping dataset...\n","โ
Dataset ready at: /content/anima_lora_training_set\n","Files found: 4\n"]}],"source":["#@markdown # Cell 4: Unzip Dataset\n","#@markdown Extracts your zipped dataset and auto-detects the inner folder if present.\n","\n","import os\n","import zipfile\n","\n","zip_path = \"/content/drive/MyDrive/lora_test.zip\" #@param {type:'string'}\n","dataset_dir = \"/content/anima_lora_training_set\"\n","\n","print(\"๐ฆ Unzipping dataset...\")\n","os.makedirs(dataset_dir, exist_ok=True)\n","\n","with zipfile.ZipFile(zip_path, 'r') as zip_ref:\n"," zip_ref.extractall(dataset_dir)\n","\n","# Auto-detect subfolder if zip had one\n","contents = os.listdir(dataset_dir)\n","if len(contents) == 1 and os.path.isdir(os.path.join(dataset_dir, contents[0])):\n"," dataset_dir = os.path.join(dataset_dir, contents[0])\n","\n","print(f\"โ
Dataset ready at: {dataset_dir}\")\n","print(f\"Files found: {len(os.listdir(dataset_dir))}\")"]},{"cell_type":"code","execution_count":18,"metadata":{"cellView":"form","colab":{"base_uri":"https://localhost:8080/","referenced_widgets":["0b5b0d9b0cb545bcacddd3b40e985c73","3018cae735df40c392b823afa06e0a40","a303e16f6cb3476897606aecfcfc673c","8268984519a1499ea48d5279c80aac07","bd3cdc62de78484a9d13feb253f770b4","441a15ff12ab49029df4384cdfb7ce87","4f7b3176a72a4bc08d1b4c844084b3ad","268a27468cb541e3a72d13e19ee48b0c","e07237b53ef244c0afb85a2103d4722f","b3ba4208332b4c9fb707517545a841dc","8f75f1e31b8d4038bdd6cef218565646"],"height":195},"id":"QfW0KpXahTzC","outputId":"efcd77c2-385b-4b80-f9b5-a64aa83d998e","executionInfo":{"status":"ok","timestamp":1775917213523,"user_tz":-120,"elapsed":435,"user":{"displayName":"No Name","userId":"10578412414437288386"}}},"outputs":[{"output_type":"stream","name":"stdout","text":["โฌ๏ธ Downloading Anima DiT model...\n","โฌ๏ธ Downloading Qwen-Image VAE...\n"]},{"output_type":"stream","name":"stderr","text":["/usr/local/lib/python3.12/dist-packages/huggingface_hub/file_download.py:982: UserWarning: `local_dir_use_symlinks` parameter is deprecated and will be ignored. The process to download files to a local folder has been updated and do not rely on symlinks anymore. You only need to pass a destination folder as`local_dir`.\n","For more details, check out https://huggingface.co/docs/huggingface_hub/main/en/guides/download#download-files-to-local-folder.\n"," warnings.warn(\n"]},{"output_type":"stream","name":"stdout","text":["โฌ๏ธ Downloading Qwen3 text encoder...\n"]},{"output_type":"display_data","data":{"text/plain":["Fetching 1 files: 0%| | 0/1 [00:00<?, ?it/s]"],"application/vnd.jupyter.widget-view+json":{"version_major":2,"version_minor":0,"model_id":"0b5b0d9b0cb545bcacddd3b40e985c73"}},"metadata":{}},{"output_type":"stream","name":"stdout","text":["โ
All Anima files downloaded to /content/anima_models/\n"]}],"source":["#@markdown # Cell 5: Download Anima DiT, VAE & Qwen3 Models\n","#@markdown Downloads the official Anima preview model + required VAE and text encoder from Hugging Face.\n","\n","from huggingface_hub import hf_hub_download, snapshot_download\n","\n","os.makedirs(\"/content/anima_models\", exist_ok=True)\n","\n","print(\"โฌ๏ธ Downloading Anima DiT model...\")\n","hf_hub_download(\n"," repo_id=\"circlestone-labs/Anima\",\n"," filename=\"split_files/diffusion_models/anima-preview2.safetensors\",\n"," local_dir=\"/content/anima_models\",\n"," token=hf_token,\n"," local_dir_use_symlinks=False\n",")\n","\n","print(\"โฌ๏ธ Downloading Qwen-Image VAE...\")\n","hf_hub_download(\n"," repo_id=\"Comfy-Org/Qwen-Image_ComfyUI\",\n"," filename=\"split_files/vae/qwen_image_vae.safetensors\",\n"," local_dir=\"/content/anima_models\",\n"," token=hf_token,\n"," local_dir_use_symlinks=False\n",")\n","\n","print(\"โฌ๏ธ Downloading Qwen3 text encoder...\")\n","snapshot_download(\n"," repo_id=\"circlestone-labs/Anima\",\n"," allow_patterns=\"split_files/text_encoders/qwen_3_06b*\",\n"," local_dir=\"/content/anima_models\",\n"," token=hf_token,\n"," local_dir_use_symlinks=False\n",")\n","\n","print(\"โ
All Anima files downloaded to /content/anima_models/\")"]},{"cell_type":"code","execution_count":19,"metadata":{"cellView":"form","colab":{"base_uri":"https://localhost:8080/"},"id":"OxGEGfo9hW-8","outputId":"fa12052d-c3a8-4acf-a9f8-293199621342","executionInfo":{"status":"ok","timestamp":1775917217838,"user_tz":-120,"elapsed":8,"user":{"displayName":"No Name","userId":"10578412414437288386"}}},"outputs":[{"output_type":"stream","name":"stdout","text":["๐ Loading DiT model to CPU...\n"," DiT loaded to CPU (~2.09B parameters)\n","๐ VAE + text encoder will stay on CPU until the training cell moves everything to VRAM\n","โ
Models pre-loaded into system RAM.\n"]}],"source":["#@markdown # Cell 6: Pre-load Models into RAM (for info only)\n","#@markdown Loads the DiT model on CPU just to show parameter count. VAE and text encoder stay on CPU until training starts.\n","\n","import torch\n","from safetensors.torch import load_file\n","\n","dit_path = \"/content/anima_models/split_files/diffusion_models/anima-preview2.safetensors\"\n","vae_path = \"/content/anima_models/split_files/vae/qwen_image_vae.safetensors\"\n","qwen3_path = \"/content/anima_models/split_files/text_encoders/qwen_3_06b_base\"\n","\n","print(\"๐ Loading DiT model to CPU...\")\n","dit_state = load_file(dit_path, device='cpu')\n","print(f\" DiT loaded to CPU (~{sum(p.numel() for p in dit_state.values()) / 1e9:.2f}B parameters)\")\n","\n","print(\"๐ VAE + text encoder will stay on CPU until the training cell moves everything to VRAM\")\n","print(\"โ
Models pre-loaded into system RAM.\")"]},{"cell_type":"code","execution_count":20,"metadata":{"colab":{"base_uri":"https://localhost:8080/"},"id":"HYlIQxproYJB","outputId":"56b8b400-02bf-4b78-8ddd-61cf7e11ed24","executionInfo":{"status":"ok","timestamp":1775917219896,"user_tz":-120,"elapsed":5,"user":{"displayName":"No Name","userId":"10578412414437288386"}}},"outputs":[{"output_type":"stream","name":"stdout","text":["๐ Dataset directory: /content/anima_lora_training_set\n"," Absolute path: /content/anima_lora_training_set\n"," Total files: 5\n"," Images found: 2\n"," Caption files (.txt) found: 2\n"," First 10 images: ['001.jpeg', '000.jpeg']\n"," First 10 captions: ['001.txt', '000.txt']\n"]}],"source":["#@markdown # Diagnostic Cell (run this FIRST โ NEW)\n","#@markdown Lists every file in your dataset folder and counts images + captions.\n","#@markdown This will tell us why the loader is exiting.\n","\n","import os\n","\n","print(f\"๐ Dataset directory: {dataset_dir}\")\n","print(f\" Absolute path: {os.path.abspath(dataset_dir)}\")\n","files = os.listdir(dataset_dir)\n","print(f\" Total files: {len(files)}\")\n","\n","images = [f for f in files if f.lower().endswith(('.png', '.jpg', '.jpeg', '.webp'))]\n","txts = [f for f in files if f.lower().endswith('.txt')]\n","\n","print(f\" Images found: {len(images)}\")\n","print(f\" Caption files (.txt) found: {len(txts)}\")\n","if len(images) > 0:\n"," print(\" First 10 images:\", images[:10])\n","if len(txts) > 0:\n"," print(\" First 10 captions:\", txts[:10])\n","else:\n"," print(\" โ ๏ธ No .txt captions found โ this is often the cause of silent exit!\")"]},{"cell_type":"code","source":["#@markdown # Cell 7: Recreate dataset_config.toml (BUCKETING DISABLED + 1024ร1024 fixed)\n","\n","dataset_config_path = os.path.join(dataset_dir, \"dataset_config.toml\")\n","\n","toml_content = f\"\"\"[general]\n","shuffle_caption = true\n","caption_extension = \".txt\"\n","enable_bucket = false # โ BUCKETING DISABLED\n","# bucket settings are ignored when enable_bucket=false\n","\n","[[datasets]]\n","resolution = 1024 # โ matches your exact image size\n","\n","[[datasets.subsets]]\n","image_dir = \"{dataset_dir}\"\n","num_repeats = 1\n","\"\"\"\n","with open(dataset_config_path, \"w\") as f:\n"," f.write(toml_content)\n","\n","print(\"โ
dataset_config.toml RECREATED โ BUCKETING DISABLED\")\n","print(f\" โ image_dir = {dataset_dir}\")\n","print(\" โ Fixed resolution = 1024ร1024 (no bucketing, no upscaling)\")"],"metadata":{"colab":{"base_uri":"https://localhost:8080/"},"id":"64MdZxdV6V40","executionInfo":{"status":"ok","timestamp":1775917575465,"user_tz":-120,"elapsed":13,"user":{"displayName":"No Name","userId":"10578412414437288386"}},"outputId":"3b414f56-668b-4252-a412-38fc693004fc"},"execution_count":25,"outputs":[{"output_type":"stream","name":"stdout","text":["โ
dataset_config.toml RECREATED โ BUCKETING DISABLED\n"," โ image_dir = /content/anima_lora_training_set\n"," โ Fixed resolution = 1024ร1024 (no bucketing, no upscaling)\n"]}]},{"cell_type":"code","source":["#@markdown # Cell 8: Editable Training Settings (โ
Qwen3 .safetensors FIXED)\n","\n","project_name = \"lora_test\"\n","output_folder = f\"/content/drive/MyDrive/anima_loras/{project_name}\"\n","\n","network_dim = 32\n","network_alpha = 16\n","learning_rate = 1e-4\n","train_batch_size = 1\n","\n","epochs = 2\n","save_every_n_epochs = 1\n","\n","os.makedirs(output_folder, exist_ok=True)\n","\n","# โ
CORRECT PATHS (Qwen3 is a .safetensors file, not a folder)\n","dit_path = \"/content/anima_models/split_files/diffusion_models/anima-preview2.safetensors\"\n","vae_path = \"/content/anima_models/split_files/vae/qwen_image_vae.safetensors\"\n","qwen3_path = \"/content/anima_models/split_files/text_encoders/qwen_3_06b_base.safetensors\" # โ this is the correct one\n","\n","dataset_config = os.path.join(dataset_dir, \"dataset_config.toml\")\n","\n","print(\"โ
Settings ready (bucketing disabled + correct Qwen3 .safetensors)\")\n","print(f\"DiT exists : {os.path.exists(dit_path)}\")\n","print(f\"VAE exists : {os.path.exists(vae_path)}\")\n","print(f\"Qwen3 file : {os.path.exists(qwen3_path)} โ should now be True\")"],"metadata":{"colab":{"base_uri":"https://localhost:8080/"},"id":"m4Slmjvb674n","executionInfo":{"status":"ok","timestamp":1775917579801,"user_tz":-120,"elapsed":7,"user":{"displayName":"No Name","userId":"10578412414437288386"}},"outputId":"8fc3b538-ab5e-4ee1-acac-75cd262610b8"},"execution_count":26,"outputs":[{"output_type":"stream","name":"stdout","text":["โ
Settings ready (bucketing disabled + correct Qwen3 .safetensors)\n","DiT exists : True\n","VAE exists : True\n","Qwen3 file : True โ should now be True\n"]}]},{"cell_type":"code","source":["#@markdown # Cell 9: ๐ Run Anima Training DIRECTLY (force full traceback + error)\n","\n","cmd = f\"\"\"\n","python -u /content/sd-scripts/anima_train_network.py \\\n"," --pretrained_model_name_or_path \"{dit_path}\" \\\n"," --vae \"{vae_path}\" \\\n"," --qwen3 \"{qwen3_path}\" \\\n"," --train_data_dir \"{dataset_dir}\" \\\n"," --resolution 1024 \\\n"," --caption_extension \".txt\" \\\n"," --shuffle_caption \\\n"," --output_dir \"{output_folder}\" \\\n"," --output_name \"{project_name}\" \\\n"," --network_module networks.lora_anima \\\n"," --network_dim {network_dim} \\\n"," --network_alpha {network_alpha} \\\n"," --learning_rate {learning_rate} \\\n"," --max_train_epochs {epochs} \\\n"," --save_every_n_epochs {save_every_n_epochs} \\\n"," --train_batch_size {train_batch_size} \\\n"," --mixed_precision \"bf16\" \\\n"," --full_bf16 \\\n"," --cache_text_encoder_outputs \\\n"," --cache_latents \\\n"," --cache_latents_to_disk \\\n"," --gradient_checkpointing \\\n"," --network_train_unet_only \\\n"," --optimizer_type \"AdamW\" \\\n"," --lr_scheduler \"cosine_with_restarts\" \\\n"," --lr_warmup_steps 0 \\\n"," --discrete_flow_shift 3 \\\n"," --vae_chunk_size 64 \\\n"," --vae_disable_cache \\\n"," --console_log_level DEBUG \\\n"," --log_config \\\n"," --max_data_loader_n_workers 0 \\\n"," --vae_batch_size 1 \\\n"," --save_model_as safetensors\n","\"\"\"\n","\n","print(\"๐ Running Anima LoRA training DIRECTLY with Python (full traceback enabled)...\")\n","print(f\" Epochs: {epochs} | Dim: {network_dim} | Batch: {train_batch_size}\")\n","print(\"โ\" * 100)\n","!{cmd}\n","print(\"โ\" * 100)\n","print(\"๐ Command finished โ check above for the REAL error or training loop!\")"],"metadata":{"colab":{"base_uri":"https://localhost:8080/"},"id":"_0lbtdLj-Otz","executionInfo":{"status":"ok","timestamp":1775918214462,"user_tz":-120,"elapsed":22274,"user":{"displayName":"No Name","userId":"10578412414437288386"}},"outputId":"d7f6bf18-7b9b-4007-dbe0-19ca34fb2ee6"},"execution_count":30,"outputs":[{"output_type":"stream","name":"stdout","text":["๐ Running Anima LoRA training DIRECTLY with Python (full traceback enabled)...\n"," Epochs: 2 | Dim: 32 | Batch: 1\n","โโโโโโโโโโโโโโโโโโโโโโโโโโโโโ๏ฟฝ๏ฟฝ๏ฟฝโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\n","2026-04-11 14:36:39.433018: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n","WARNING: All log messages before absl::InitializeLog() is called are written to STDERR\n","E0000 00:00:1775918199.453778 22318 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n","E0000 00:00:1775918199.460036 22318 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n","W0000 00:00:1775918199.476289 22318 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\n","W0000 00:00:1775918199.476352 22318 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\n","W0000 00:00:1775918199.476360 22318 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\n","W0000 00:00:1775918199.476364 22318 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\n","2026-04-11 14:36:39.481455: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\n","To enable the following instructions: AVX2 AVX512F FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n","\u001b[2;36m2026-04-11 14:36:49\u001b[0m\u001b[2;36m \u001b[0m\u001b[34mINFO \u001b[0m Using DreamBooth method. \u001b]8;id=259606;file:///content/sd-scripts/train_network.py\u001b\\\u001b[2mtrain_network.py\u001b[0m\u001b]8;;\u001b\\\u001b[2m:\u001b[0m\u001b]8;id=654001;file:///content/sd-scripts/train_network.py#513\u001b\\\u001b[2m513\u001b[0m\u001b]8;;\u001b\\\n","\u001b[2;36m \u001b[0m\u001b[2;36m \u001b[0m\u001b[34mINFO \u001b[0m prepare images. \u001b]8;id=825465;file:///content/sd-scripts/library/train_util.py\u001b\\\u001b[2mtrain_util.py\u001b[0m\u001b]8;;\u001b\\\u001b[2m:\u001b[0m\u001b]8;id=689569;file:///content/sd-scripts/library/train_util.py#2141\u001b\\\u001b[2m2141\u001b[0m\u001b]8;;\u001b\\\n","\u001b[2;36m \u001b[0m\u001b[2;36m \u001b[0m\u001b[34mINFO \u001b[0m \u001b[1;36m0\u001b[0m train images with repeats. \u001b]8;id=337480;file:///content/sd-scripts/library/train_util.py\u001b\\\u001b[2mtrain_util.py\u001b[0m\u001b]8;;\u001b\\\u001b[2m:\u001b[0m\u001b]8;id=135783;file:///content/sd-scripts/library/train_util.py#2187\u001b\\\u001b[2m2187\u001b[0m\u001b]8;;\u001b\\\n","\u001b[2;36m \u001b[0m\u001b[2;36m \u001b[0m\u001b[34mINFO \u001b[0m \u001b[1;36m0\u001b[0m reg images with repeats. \u001b]8;id=404396;file:///content/sd-scripts/library/train_util.py\u001b\\\u001b[2mtrain_util.py\u001b[0m\u001b]8;;\u001b\\\u001b[2m:\u001b[0m\u001b]8;id=760020;file:///content/sd-scripts/library/train_util.py#2191\u001b\\\u001b[2m2191\u001b[0m\u001b]8;;\u001b\\\n","\u001b[2;36m \u001b[0m\u001b[2;36m \u001b[0m\u001b[33mWARNING \u001b[0m no regularization images \u001b[35m/\u001b[0m \u001b]8;id=638969;file:///content/sd-scripts/library/train_util.py\u001b\\\u001b[2mtrain_util.py\u001b[0m\u001b]8;;\u001b\\\u001b[2m:\u001b[0m\u001b]8;id=260755;file:///content/sd-scripts/library/train_util.py#2196\u001b\\\u001b[2m2196\u001b[0m\u001b]8;;\u001b\\\n","\u001b[2;36m \u001b[0m ๆญฃๅๅ็ปๅใ่ฆใคใใใพใใใงใใ \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m\u001b[2;36m \u001b[0m\u001b[34mINFO \u001b[0m \u001b[1m[\u001b[0mDataset \u001b[1;36m0\u001b[0m\u001b[1m]\u001b[0m \u001b]8;id=761068;file:///content/sd-scripts/library/config_util.py\u001b\\\u001b[2mconfig_util.py\u001b[0m\u001b]8;;\u001b\\\u001b[2m:\u001b[0m\u001b]8;id=591107;file:///content/sd-scripts/library/config_util.py#584\u001b\\\u001b[2m584\u001b[0m\u001b]8;;\u001b\\\n","\u001b[2;36m \u001b[0m batch_size: \u001b[1;36m1\u001b[0m \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m resolution: \u001b[1m(\u001b[0m\u001b[1;36m1024\u001b[0m, \u001b[1;36m1024\u001b[0m\u001b[1m)\u001b[0m \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m skip_image_resolution: \u001b[3;35mNone\u001b[0m \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m resize_interpolation: \u001b[3;35mNone\u001b[0m \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m enable_bucket: \u001b[3;91mFalse\u001b[0m \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m\u001b[2;36m \u001b[0m\u001b[34mINFO \u001b[0m \u001b[1m[\u001b[0mPrepare dataset \u001b[1;36m0\u001b[0m\u001b[1m]\u001b[0m \u001b]8;id=237856;file:///content/sd-scripts/library/config_util.py\u001b\\\u001b[2mconfig_util.py\u001b[0m\u001b]8;;\u001b\\\u001b[2m:\u001b[0m\u001b]8;id=298464;file:///content/sd-scripts/library/config_util.py#596\u001b\\\u001b[2m596\u001b[0m\u001b]8;;\u001b\\\n","\u001b[2;36m \u001b[0m\u001b[2;36m \u001b[0m\u001b[34mINFO \u001b[0m loading image sizes. \u001b]8;id=657189;file:///content/sd-scripts/library/train_util.py\u001b\\\u001b[2mtrain_util.py\u001b[0m\u001b]8;;\u001b\\\u001b[2m:\u001b[0m\u001b]8;id=545212;file:///content/sd-scripts/library/train_util.py#995\u001b\\\u001b[2m995\u001b[0m\u001b]8;;\u001b\\\n","0it [00:00, ?it/s]\n","\u001b[2;36m \u001b[0m\u001b[2;36m \u001b[0m\u001b[34mINFO \u001b[0m prepare dataset \u001b]8;id=878742;file:///content/sd-scripts/library/train_util.py\u001b\\\u001b[2mtrain_util.py\u001b[0m\u001b]8;;\u001b\\\u001b[2m:\u001b[0m\u001b]8;id=553022;file:///content/sd-scripts/library/train_util.py#1020\u001b\\\u001b[2m1020\u001b[0m\u001b]8;;\u001b\\\n","\u001b[2;36m \u001b[0m\u001b[2;36m \u001b[0m\u001b[1;31mERROR \u001b[0m No data found. Please verify \u001b]8;id=42450;file:///content/sd-scripts/train_network.py\u001b\\\u001b[2mtrain_network.py\u001b[0m\u001b]8;;\u001b\\\u001b[2m:\u001b[0m\u001b]8;id=271493;file:///content/sd-scripts/train_network.py#559\u001b\\\u001b[2m559\u001b[0m\u001b]8;;\u001b\\\n","\u001b[2;36m \u001b[0m arguments \u001b[1m(\u001b[0mtrain_data_dir must \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m be the parent of folders with \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m images\u001b[1m)\u001b[0m \u001b[35m/\u001b[0m \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m ็ปๅใใใใพใใใๅผๆฐๆๅฎใ็ขบ \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m ่ชใใฆใใ ใใ๏ผtrain_data_dir \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m ใซใฏ็ปๅใใใใใฉใซใใงใฏใชใ \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m ใ็ปๅใใใใใฉใซใใฎ่ฆชใใฉใซ \u001b[2m \u001b[0m\n","\u001b[2;36m \u001b[0m ใใๆๅฎใใๅฟ
่ฆใใใใพใ๏ผ \u001b[2m \u001b[0m\n","โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ\n","๐ Command finished โ check above for the REAL error or training loop!\n"]}]}],"metadata":{"accelerator":"GPU","colab":{"gpuType":"T4","provenance":[],"authorship_tag":"ABX9TyPfqA0myZuBaeipqiAW3nbc"},"kernelspec":{"display_name":"Python 3","name":"python3"},"language_info":{"name":"python"},"widgets":{"application/vnd.jupyter.widget-state+json":{"0b5b0d9b0cb545bcacddd3b40e985c73":{"model_module":"@jupyter-widgets/controls","model_name":"HBoxModel","model_module_version":"1.5.0","state":{"_dom_classes":[],"_model_module":"@jupyter-widgets/controls","_model_module_version":"1.5.0","_model_name":"HBoxModel","_view_count":null,"_view_module":"@jupyter-widgets/controls","_view_module_version":"1.5.0","_view_name":"HBoxView","box_style":"","children":["IPY_MODEL_3018cae735df40c392b823afa06e0a40","IPY_MODEL_a303e16f6cb3476897606aecfcfc673c","IPY_MODEL_8268984519a1499ea48d5279c80aac07"],"layout":"IPY_MODEL_bd3cdc62de78484a9d13feb253f770b4"}},"3018cae735df40c392b823afa06e0a40":{"model_module":"@jupyter-widgets/controls","model_name":"HTMLModel","model_module_version":"1.5.0","state":{"_dom_classes":[],"_model_module":"@jupyter-widgets/controls","_model_module_version":"1.5.0","_model_name":"HTMLModel","_view_count":null,"_view_module":"@jupyter-widgets/controls","_view_module_version":"1.5.0","_view_name":"HTMLView","description":"","description_tooltip":null,"layout":"IPY_MODEL_441a15ff12ab49029df4384cdfb7ce87","placeholder":"โ","style":"IPY_MODEL_4f7b3176a72a4bc08d1b4c844084b3ad","value":"Fetchingโ1โfiles:โ100%"}},"a303e16f6cb3476897606aecfcfc673c":{"model_module":"@jupyter-widgets/controls","model_name":"FloatProgressModel","model_module_version":"1.5.0","state":{"_dom_classes":[],"_model_module":"@jupyter-widgets/controls","_model_module_version":"1.5.0","_model_name":"FloatProgressModel","_view_count":null,"_view_module":"@jupyter-widgets/controls","_view_module_version":"1.5.0","_view_name":"ProgressView","bar_style":"success","description":"","description_tooltip":null,"layout":"IPY_MODEL_268a27468cb541e3a72d13e19ee48b0c","max":1,"min":0,"orientation":"horizontal","style":"IPY_MODEL_e07237b53ef244c0afb85a2103d4722f","value":1}},"8268984519a1499ea48d5279c80aac07":{"model_module":"@jupyter-widgets/controls","model_name":"HTMLModel","model_module_version":"1.5.0","state":{"_dom_classes":[],"_model_module":"@jupyter-widgets/controls","_model_module_version":"1.5.0","_model_name":"HTMLModel","_view_count":null,"_view_module":"@jupyter-widgets/controls","_view_module_version":"1.5.0","_view_name":"HTMLView","description":"","description_tooltip":null,"layout":"IPY_MODEL_b3ba4208332b4c9fb707517545a841dc","placeholder":"โ","style":"IPY_MODEL_8f75f1e31b8d4038bdd6cef218565646","value":"โ1/1โ[00:00<00:00,โ65.70it/s]"}},"bd3cdc62de78484a9d13feb253f770b4":{"model_module":"@jupyter-widgets/base","model_name":"LayoutModel","model_module_version":"1.2.0","state":{"_model_module":"@jupyter-widgets/base","_model_module_version":"1.2.0","_model_name":"LayoutModel","_view_count":null,"_view_module":"@jupyter-widgets/base","_view_module_version":"1.2.0","_view_name":"LayoutView","align_content":null,"align_items":null,"align_self":null,"border":null,"bottom":null,"display":null,"flex":null,"flex_flow":null,"grid_area":null,"grid_auto_columns":null,"grid_auto_flow":null,"grid_auto_rows":null,"grid_column":null,"grid_gap":null,"grid_row":null,"grid_template_areas":null,"grid_template_columns":null,"grid_template_rows":null,"height":null,"justify_content":null,"justify_items":null,"left":null,"margin":null,"max_height":null,"max_width":null,"min_height":null,"min_width":null,"object_fit":null,"object_position":null,"order":null,"overflow":null,"overflow_x":null,"overflow_y":null,"padding":null,"right":null,"top":null,"visibility":null,"width":null}},"441a15ff12ab49029df4384cdfb7ce87":{"model_module":"@jupyter-widgets/base","model_name":"LayoutModel","model_module_version":"1.2.0","state":{"_model_module":"@jupyter-widgets/base","_model_module_version":"1.2.0","_model_name":"LayoutModel","_view_count":null,"_view_module":"@jupyter-widgets/base","_view_module_version":"1.2.0","_view_name":"LayoutView","align_content":null,"align_items":null,"align_self":null,"border":null,"bottom":null,"display":null,"flex":null,"flex_flow":null,"grid_area":null,"grid_auto_columns":null,"grid_auto_flow":null,"grid_auto_rows":null,"grid_column":null,"grid_gap":null,"grid_row":null,"grid_template_areas":null,"grid_template_columns":null,"grid_template_rows":null,"height":null,"justify_content":null,"justify_items":null,"left":null,"margin":null,"max_height":null,"max_width":null,"min_height":null,"min_width":null,"object_fit":null,"object_position":null,"order":null,"overflow":null,"overflow_x":null,"overflow_y":null,"padding":null,"right":null,"top":null,"visibility":null,"width":null}},"4f7b3176a72a4bc08d1b4c844084b3ad":{"model_module":"@jupyter-widgets/controls","model_name":"DescriptionStyleModel","model_module_version":"1.5.0","state":{"_model_module":"@jupyter-widgets/controls","_model_module_version":"1.5.0","_model_name":"DescriptionStyleModel","_view_count":null,"_view_module":"@jupyter-widgets/base","_view_module_version":"1.2.0","_view_name":"StyleView","description_width":""}},"268a27468cb541e3a72d13e19ee48b0c":{"model_module":"@jupyter-widgets/base","model_name":"LayoutModel","model_module_version":"1.2.0","state":{"_model_module":"@jupyter-widgets/base","_model_module_version":"1.2.0","_model_name":"LayoutModel","_view_count":null,"_view_module":"@jupyter-widgets/base","_view_module_version":"1.2.0","_view_name":"LayoutView","align_content":null,"align_items":null,"align_self":null,"border":null,"bottom":null,"display":null,"flex":null,"flex_flow":null,"grid_area":null,"grid_auto_columns":null,"grid_auto_flow":null,"grid_auto_rows":null,"grid_column":null,"grid_gap":null,"grid_row":null,"grid_template_areas":null,"grid_template_columns":null,"grid_template_rows":null,"height":null,"justify_content":null,"justify_items":null,"left":null,"margin":null,"max_height":null,"max_width":null,"min_height":null,"min_width":null,"object_fit":null,"object_position":null,"order":null,"overflow":null,"overflow_x":null,"overflow_y":null,"padding":null,"right":null,"top":null,"visibility":null,"width":null}},"e07237b53ef244c0afb85a2103d4722f":{"model_module":"@jupyter-widgets/controls","model_name":"ProgressStyleModel","model_module_version":"1.5.0","state":{"_model_module":"@jupyter-widgets/controls","_model_module_version":"1.5.0","_model_name":"ProgressStyleModel","_view_count":null,"_view_module":"@jupyter-widgets/base","_view_module_version":"1.2.0","_view_name":"StyleView","bar_color":null,"description_width":""}},"b3ba4208332b4c9fb707517545a841dc":{"model_module":"@jupyter-widgets/base","model_name":"LayoutModel","model_module_version":"1.2.0","state":{"_model_module":"@jupyter-widgets/base","_model_module_version":"1.2.0","_model_name":"LayoutModel","_view_count":null,"_view_module":"@jupyter-widgets/base","_view_module_version":"1.2.0","_view_name":"LayoutView","align_content":null,"align_items":null,"align_self":null,"border":null,"bottom":null,"display":null,"flex":null,"flex_flow":null,"grid_area":null,"grid_auto_columns":null,"grid_auto_flow":null,"grid_auto_rows":null,"grid_column":null,"grid_gap":null,"grid_row":null,"grid_template_areas":null,"grid_template_columns":null,"grid_template_rows":null,"height":null,"justify_content":null,"justify_items":null,"left":null,"margin":null,"max_height":null,"max_width":null,"min_height":null,"min_width":null,"object_fit":null,"object_position":null,"order":null,"overflow":null,"overflow_x":null,"overflow_y":null,"padding":null,"right":null,"top":null,"visibility":null,"width":null}},"8f75f1e31b8d4038bdd6cef218565646":{"model_module":"@jupyter-widgets/controls","model_name":"DescriptionStyleModel","model_module_version":"1.5.0","state":{"_model_module":"@jupyter-widgets/controls","_model_module_version":"1.5.0","_model_name":"DescriptionStyleModel","_view_count":null,"_view_module":"@jupyter-widgets/base","_view_module_version":"1.2.0","_view_name":"StyleView","description_width":""}}}}},"nbformat":4,"nbformat_minor":0}
|