diff --git "a/.ipynb_checkpoints/unet_stas-checkpoint.ipynb" "b/.ipynb_checkpoints/unet_stas-checkpoint.ipynb" new file mode 100644--- /dev/null +++ "b/.ipynb_checkpoints/unet_stas-checkpoint.ipynb" @@ -0,0 +1,1758 @@ +{ + "cells": [ + { + "cell_type": "code", + "execution_count": 11, + "id": "82ca7882-410c-4067-863a-07838d485f6a", + "metadata": {}, + "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "test unet\n", + "Количество параметров: 1393037440\n", + "Output shape: torch.Size([1, 128, 60, 48])\n", + "Output shape: torch.Size([1, 128, 60, 48])\n" + ] + } + ], + "source": [ + "config_sdxs = {\n", + " # === Основные размеры и каналы ===\n", + " \"in_channels\": 128, # Количество входных каналов (совместимость с 16-канальным VAE)\n", + " \"out_channels\": 128, # Количество выходных каналов (симметрично in_channels)\n", + " \"center_input_sample\": False, # Отключение центрирования входных данных (стандарт для диффузионных моделей)\n", + " \"flip_sin_to_cos\": True, # Автоматическое преобразование sin/cos в эмбеддингах времени (для стабильности)\n", + " \"freq_shift\": 0, # Сдвиг частоты (0 - стандартное значение для частотных эмбеддингов)\n", + "\n", + " # === Архитектура блоков ===\n", + " \"down_block_types\": [ # Типы блоков энкодера (иерархия обработки):\n", + " \"DownBlock2D\",\n", + " \"CrossAttnDownBlock2D\",\n", + " \"CrossAttnDownBlock2D\",\n", + " ],\n", + " \"mid_block_type\": \"UNetMidBlock2DCrossAttn\", # Центральный блок с cross-attention (бутылочное горлышко сети)\n", + " \"up_block_types\": [ # Типы блоков декодера (восстановление изображения):\n", + " \"CrossAttnUpBlock2D\",\n", + " \"CrossAttnUpBlock2D\",\n", + " \"UpBlock2D\",\n", + " ],\n", + " \"only_cross_attention\": False, # Использование как cross-attention, так и self-attention\n", + "\n", + " # === Конфигурация каналов ===\n", + " \"block_out_channels\": [512, 1024, 1280], \n", + " \"layers_per_block\": 2, # Число слоев в блоках\n", + " \"downsample_padding\": 1, # Паддинг при уменьшении разрешения\n", + " \"mid_block_scale_factor\": 1.0, # Усиление сигнала в центральном блоке\n", + "\n", + " # === Нормализация ===\n", + " \"norm_num_groups\": 32, # Число групп для GroupNorm (оптимально для стабильности)\n", + " \"norm_eps\": 1e-05, # Эпсилон для нормализации (стандартное значение)\n", + "\n", + " # === Cross-Attention ===\n", + " \"cross_attention_dim\": 768, # Размерность текстовых эмбеддинго\n", + " \n", + " \"transformer_layers_per_block\": 3, # Число трансформерных слоев (уменьшение с глубиной)\n", + " \"attention_head_dim\": [16,16,20], # Размерность головы внимания \n", + " \"dual_cross_attention\": False, # Отключение двойного внимания (упрощение архитектуры)\n", + " \"use_linear_projection\": False, # Изменено на True для лучшей организации памяти\n", + "\n", + " # === ResNet Блоки ===\n", + " \"resnet_time_scale_shift\": \"default\", # Способ интеграции временных эмбеддингов\n", + " \"resnet_skip_time_act\": False, # Отключение активации в skip-соединениях\n", + " \"resnet_out_scale_factor\": 1.0, # Коэффициент масштабирования выхода ResNet\n", + "\n", + " # === Временные эмбеддинги ===\n", + " \"time_embedding_type\": \"positional\", # Тип временных эмбеддингов (стандартный подход)\n", + "\n", + " # === Свертки ===\n", + " \"conv_in_kernel\": 3, # Ядро входной свертки (баланс между рецептивным полем и параметрами)\n", + " \"conv_out_kernel\": 3, # Ядро выходной свертки (симметрично входной)\n", + "}\n", + "\n", + "if 1:\n", + " checkpoint_path = \"tmp\"#\"sdxs\"\n", + " import torch\n", + " from diffusers import UNet2DConditionModel\n", + " print(\"test unet\")\n", + " new_unet = UNet2DConditionModel(**config_sdxs).to(\"cuda\", dtype=torch.float16)\n", + "\n", + " assert all(ch % 32 == 0 for ch in new_unet.config[\"block_out_channels\"]), \"Каналы должны быть кратны 32\"\n", + " num_params = sum(p.numel() for p in new_unet.parameters())\n", + " print(f\"Количество параметров: {num_params}\")\n", + "\n", + " # Генерация тестового латента (640x512 в latent space)\n", + " test_latent = torch.randn(1, 128, 60, 48).to(\"cuda\", dtype=torch.float16) # 60x48 ≈ 512px\n", + " timesteps = torch.tensor([1]).to(\"cuda\", dtype=torch.float16)\n", + " encoder_hidden_states = torch.randn(1, 77, 768).to(\"cuda\", dtype=torch.float16)\n", + " \n", + " with torch.no_grad():\n", + " output = new_unet(\n", + " test_latent, \n", + " timesteps, \n", + " encoder_hidden_states\n", + " ).sample\n", + " \n", + " print(f\"Output shape: {output.shape}\") \n", + " new_unet.save_pretrained(checkpoint_path)\n", + " #print(new_unet)\n", + " del new_unet\n", + " torch.cuda.empty_cache()\n", + " print(f\"Output shape: {output.shape}\") \n", + " # Количество параметров: 1601774976" + ] + }, + { + "cell_type": "code", + "execution_count": 12, + "id": "f980bb1a-9859-44c2-a2df-ff1b073bf435", + "metadata": {}, + "outputs": [ + { + "name": "stderr", + "output_type": "stream", + "text": [ + "Перенос весов: 100%|██████████| 904/904 [00:00<00:00, 112132.57it/s]\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "✗ Несовпадение размеров: conv_in.weight (torch.Size([1024, 128, 1, 1])) -> conv_in.weight (torch.Size([512, 128, 3, 3]))\n", + "✗ Несовпадение размеров: conv_in.bias (torch.Size([1024])) -> conv_in.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: time_embedding.linear_1.weight (torch.Size([4096, 1024])) -> time_embedding.linear_1.weight (torch.Size([2048, 512]))\n", + "✗ Несовпадение размеров: time_embedding.linear_1.bias (torch.Size([4096])) -> time_embedding.linear_1.bias (torch.Size([2048]))\n", + "✗ Несовпадение размеров: time_embedding.linear_2.weight (torch.Size([4096, 4096])) -> time_embedding.linear_2.weight (torch.Size([2048, 2048]))\n", + "✗ Несовпадение размеров: time_embedding.linear_2.bias (torch.Size([4096])) -> time_embedding.linear_2.bias (torch.Size([2048]))\n", + "✗ Несовпадение размеров: down_blocks.0.resnets.0.norm1.weight (torch.Size([1024])) -> down_blocks.0.resnets.0.norm1.weight (torch.Size([512]))\n", + "✗ Несовпадение размеров: down_blocks.0.resnets.0.norm1.bias (torch.Size([1024])) -> down_blocks.0.resnets.0.norm1.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: down_blocks.0.resnets.0.conv1.weight (torch.Size([1024, 1024, 3, 3])) -> down_blocks.0.resnets.0.conv1.weight (torch.Size([512, 512, 3, 3]))\n", + "✗ Несовпадение размеров: down_blocks.0.resnets.0.conv1.bias (torch.Size([1024])) -> down_blocks.0.resnets.0.conv1.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: down_blocks.0.resnets.0.time_emb_proj.weight (torch.Size([1024, 4096])) -> down_blocks.0.resnets.0.time_emb_proj.weight (torch.Size([512, 2048]))\n", + "✗ Несовпадение размеров: down_blocks.0.resnets.0.time_emb_proj.bias (torch.Size([1024])) -> down_blocks.0.resnets.0.time_emb_proj.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: down_blocks.0.resnets.0.norm2.weight (torch.Size([1024])) -> down_blocks.0.resnets.0.norm2.weight (torch.Size([512]))\n", + "✗ Несовпадение размеров: down_blocks.0.resnets.0.norm2.bias (torch.Size([1024])) -> down_blocks.0.resnets.0.norm2.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: down_blocks.0.resnets.0.conv2.weight (torch.Size([1024, 1024, 3, 3])) -> down_blocks.0.resnets.0.conv2.weight (torch.Size([512, 512, 3, 3]))\n", + "✗ Несовпадение размеров: down_blocks.0.resnets.0.conv2.bias (torch.Size([1024])) -> down_blocks.0.resnets.0.conv2.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: down_blocks.0.downsamplers.0.conv.weight (torch.Size([1024, 1024, 3, 3])) -> down_blocks.0.downsamplers.0.conv.weight (torch.Size([512, 512, 3, 3]))\n", + "✗ Несовпадение размеров: down_blocks.0.downsamplers.0.conv.bias (torch.Size([1024])) -> down_blocks.0.downsamplers.0.conv.bias (torch.Size([512]))\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.norm1.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.norm1.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.attn1.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.attn1.to_k.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.attn1.to_v.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.attn1.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.attn1.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.norm2.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.norm2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.attn2.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.attn2.to_k.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.attn2.to_v.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.attn2.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.attn2.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.norm3.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.norm3.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.ff.net.0.proj.weight -> torch.Size([8192, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.ff.net.0.proj.bias -> torch.Size([8192])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.ff.net.2.weight -> torch.Size([1024, 4096])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.3.ff.net.2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.norm1.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.norm1.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.attn1.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.attn1.to_k.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.attn1.to_v.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.attn1.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.attn1.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.norm2.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.norm2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.attn2.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.attn2.to_k.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.attn2.to_v.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.attn2.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.attn2.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.norm3.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.norm3.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.ff.net.0.proj.weight -> torch.Size([8192, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.ff.net.0.proj.bias -> torch.Size([8192])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.ff.net.2.weight -> torch.Size([1024, 4096])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.4.ff.net.2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.norm1.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.norm1.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.attn1.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.attn1.to_k.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.attn1.to_v.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.attn1.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.attn1.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.norm2.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.norm2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.attn2.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.attn2.to_k.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.attn2.to_v.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.attn2.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.attn2.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.norm3.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.norm3.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.ff.net.0.proj.weight -> torch.Size([8192, 1024])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.ff.net.0.proj.bias -> torch.Size([8192])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.ff.net.2.weight -> torch.Size([1024, 4096])\n", + "? Ключ не найден в новой модели: down_blocks.1.attentions.0.transformer_blocks.5.ff.net.2.bias -> torch.Size([1024])\n", + "✗ Несовпадение размеров: down_blocks.1.resnets.0.norm1.weight (torch.Size([1024])) -> down_blocks.1.resnets.0.norm1.weight (torch.Size([512]))\n", + "✗ Несовпадение размеров: down_blocks.1.resnets.0.norm1.bias (torch.Size([1024])) -> down_blocks.1.resnets.0.norm1.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: down_blocks.1.resnets.0.conv1.weight (torch.Size([1024, 1024, 3, 3])) -> down_blocks.1.resnets.0.conv1.weight (torch.Size([1024, 512, 3, 3]))\n", + "✗ Несовпадение размеров: down_blocks.1.resnets.0.time_emb_proj.weight (torch.Size([1024, 4096])) -> down_blocks.1.resnets.0.time_emb_proj.weight (torch.Size([1024, 2048]))\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.norm1.weight -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.norm1.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.attn1.to_q.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.attn1.to_k.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.attn1.to_v.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.attn1.to_out.0.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.attn1.to_out.0.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.norm2.weight -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.norm2.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.attn2.to_q.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.attn2.to_k.weight -> torch.Size([1280, 768])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.attn2.to_v.weight -> torch.Size([1280, 768])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.attn2.to_out.0.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.attn2.to_out.0.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.norm3.weight -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.norm3.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.ff.net.0.proj.weight -> torch.Size([10240, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.ff.net.0.proj.bias -> torch.Size([10240])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.ff.net.2.weight -> torch.Size([1280, 5120])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.3.ff.net.2.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.norm1.weight -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.norm1.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.attn1.to_q.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.attn1.to_k.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.attn1.to_v.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.attn1.to_out.0.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.attn1.to_out.0.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.norm2.weight -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.norm2.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.attn2.to_q.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.attn2.to_k.weight -> torch.Size([1280, 768])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.attn2.to_v.weight -> torch.Size([1280, 768])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.attn2.to_out.0.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.attn2.to_out.0.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.norm3.weight -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.norm3.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.ff.net.0.proj.weight -> torch.Size([10240, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.ff.net.0.proj.bias -> torch.Size([10240])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.ff.net.2.weight -> torch.Size([1280, 5120])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.4.ff.net.2.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.norm1.weight -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.norm1.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.attn1.to_q.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.attn1.to_k.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.attn1.to_v.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.attn1.to_out.0.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.attn1.to_out.0.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.norm2.weight -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.norm2.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.attn2.to_q.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.attn2.to_k.weight -> torch.Size([1280, 768])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.attn2.to_v.weight -> torch.Size([1280, 768])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.attn2.to_out.0.weight -> torch.Size([1280, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.attn2.to_out.0.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.norm3.weight -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.norm3.bias -> torch.Size([1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.ff.net.0.proj.weight -> torch.Size([10240, 1280])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.ff.net.0.proj.bias -> torch.Size([10240])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.ff.net.2.weight -> torch.Size([1280, 5120])\n", + "? Ключ не найден в новой модели: down_blocks.2.attentions.0.transformer_blocks.5.ff.net.2.bias -> torch.Size([1280])\n", + "✗ Несовпадение размеров: down_blocks.2.resnets.0.time_emb_proj.weight (torch.Size([1280, 4096])) -> down_blocks.2.resnets.0.time_emb_proj.weight (torch.Size([1280, 2048]))\n", + "✗ Несовпадение размеров: up_blocks.0.resnets.0.time_emb_proj.weight (torch.Size([1280, 4096])) -> up_blocks.0.resnets.0.time_emb_proj.weight (torch.Size([1280, 2048]))\n", + "✗ Несовпадение размеров: up_blocks.0.resnets.1.norm1.weight (torch.Size([2304])) -> up_blocks.0.resnets.1.norm1.weight (torch.Size([2560]))\n", + "✗ Несовпадение размеров: up_blocks.0.resnets.1.norm1.bias (torch.Size([2304])) -> up_blocks.0.resnets.1.norm1.bias (torch.Size([2560]))\n", + "✗ Несовпадение размеров: up_blocks.0.resnets.1.conv1.weight (torch.Size([1280, 2304, 3, 3])) -> up_blocks.0.resnets.1.conv1.weight (torch.Size([1280, 2560, 3, 3]))\n", + "✗ Несовпадение размеров: up_blocks.0.resnets.1.time_emb_proj.weight (torch.Size([1280, 4096])) -> up_blocks.0.resnets.1.time_emb_proj.weight (torch.Size([1280, 2048]))\n", + "✗ Несовпадение размеров: up_blocks.0.resnets.1.conv_shortcut.weight (torch.Size([1280, 2304, 1, 1])) -> up_blocks.0.resnets.1.conv_shortcut.weight (torch.Size([1280, 2560, 1, 1]))\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.norm1.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.norm1.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.attn1.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.attn1.to_k.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.attn1.to_v.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.attn1.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.attn1.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.norm2.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.norm2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.attn2.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.attn2.to_k.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.attn2.to_v.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.attn2.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.attn2.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.norm3.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.norm3.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.ff.net.0.proj.weight -> torch.Size([8192, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.ff.net.0.proj.bias -> torch.Size([8192])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.ff.net.2.weight -> torch.Size([1024, 4096])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.3.ff.net.2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.norm1.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.norm1.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.attn1.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.attn1.to_k.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.attn1.to_v.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.attn1.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.attn1.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.norm2.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.norm2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.attn2.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.attn2.to_k.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.attn2.to_v.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.attn2.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.attn2.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.norm3.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.norm3.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.ff.net.0.proj.weight -> torch.Size([8192, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.ff.net.0.proj.bias -> torch.Size([8192])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.ff.net.2.weight -> torch.Size([1024, 4096])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.4.ff.net.2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.norm1.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.norm1.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.attn1.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.attn1.to_k.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.attn1.to_v.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.attn1.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.attn1.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.norm2.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.norm2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.attn2.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.attn2.to_k.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.attn2.to_v.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.attn2.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.attn2.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.norm3.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.norm3.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.ff.net.0.proj.weight -> torch.Size([8192, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.ff.net.0.proj.bias -> torch.Size([8192])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.ff.net.2.weight -> torch.Size([1024, 4096])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.0.transformer_blocks.5.ff.net.2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.norm1.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.norm1.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.attn1.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.attn1.to_k.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.attn1.to_v.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.attn1.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.attn1.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.norm2.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.norm2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.attn2.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.attn2.to_k.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.attn2.to_v.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.attn2.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.attn2.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.norm3.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.norm3.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.ff.net.0.proj.weight -> torch.Size([8192, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.ff.net.0.proj.bias -> torch.Size([8192])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.ff.net.2.weight -> torch.Size([1024, 4096])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.3.ff.net.2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.norm1.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.norm1.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.attn1.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.attn1.to_k.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.attn1.to_v.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.attn1.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.attn1.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.norm2.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.norm2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.attn2.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.attn2.to_k.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.attn2.to_v.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.attn2.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.attn2.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.norm3.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.norm3.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.ff.net.0.proj.weight -> torch.Size([8192, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.ff.net.0.proj.bias -> torch.Size([8192])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.ff.net.2.weight -> torch.Size([1024, 4096])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.4.ff.net.2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.norm1.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.norm1.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.attn1.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.attn1.to_k.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.attn1.to_v.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.attn1.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.attn1.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.norm2.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.norm2.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.attn2.to_q.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.attn2.to_k.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.attn2.to_v.weight -> torch.Size([1024, 768])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.attn2.to_out.0.weight -> torch.Size([1024, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.attn2.to_out.0.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.norm3.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.norm3.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.ff.net.0.proj.weight -> torch.Size([8192, 1024])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.ff.net.0.proj.bias -> torch.Size([8192])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.ff.net.2.weight -> torch.Size([1024, 4096])\n", + "? Ключ не найден в новой модели: up_blocks.1.attentions.1.transformer_blocks.5.ff.net.2.bias -> torch.Size([1024])\n", + "✗ Несовпадение размеров: up_blocks.1.resnets.0.time_emb_proj.weight (torch.Size([1024, 4096])) -> up_blocks.1.resnets.0.time_emb_proj.weight (torch.Size([1024, 2048]))\n", + "✗ Несовпадение размеров: up_blocks.1.resnets.1.time_emb_proj.weight (torch.Size([1024, 4096])) -> up_blocks.1.resnets.1.time_emb_proj.weight (torch.Size([1024, 2048]))\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.norm.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.norm.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.proj_in.weight -> torch.Size([1020, 1024, 1, 1])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.proj_in.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.0.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.1.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.2.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.3.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.4.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой мо��ели: up_blocks.2.attentions.0.transformer_blocks.4.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.transformer_blocks.5.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.proj_out.weight -> torch.Size([1024, 1020, 1, 1])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.0.proj_out.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.norm.weight -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.norm.bias -> torch.Size([1024])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.proj_in.weight -> torch.Size([1020, 1024, 1, 1])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.proj_in.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.1.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.2.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.3.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.4.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.norm1.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.norm1.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.attn1.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.attn1.to_k.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.attn1.to_v.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.attn1.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.attn1.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.norm2.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.norm2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.attn2.to_q.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.attn2.to_k.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.attn2.to_v.weight -> torch.Size([1020, 768])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.attn2.to_out.0.weight -> torch.Size([1020, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.attn2.to_out.0.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.norm3.weight -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.norm3.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.ff.net.0.proj.weight -> torch.Size([8160, 1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.ff.net.0.proj.bias -> torch.Size([8160])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.ff.net.2.weight -> torch.Size([1020, 4080])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.transformer_blocks.5.ff.net.2.bias -> torch.Size([1020])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.proj_out.weight -> torch.Size([1024, 1020, 1, 1])\n", + "? Ключ не найден в новой модели: up_blocks.2.attentions.1.proj_out.bias -> torch.Size([1024])\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.norm1.weight (torch.Size([2048])) -> up_blocks.2.resnets.0.norm1.weight (torch.Size([1536]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.norm1.bias (torch.Size([2048])) -> up_blocks.2.resnets.0.norm1.bias (torch.Size([1536]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.conv1.weight (torch.Size([1024, 2048, 3, 3])) -> up_blocks.2.resnets.0.conv1.weight (torch.Size([512, 1536, 3, 3]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.conv1.bias (torch.Size([1024])) -> up_blocks.2.resnets.0.conv1.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.time_emb_proj.weight (torch.Size([1024, 4096])) -> up_blocks.2.resnets.0.time_emb_proj.weight (torch.Size([512, 2048]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.time_emb_proj.bias (torch.Size([1024])) -> up_blocks.2.resnets.0.time_emb_proj.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.norm2.weight (torch.Size([1024])) -> up_blocks.2.resnets.0.norm2.weight (torch.Size([512]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.norm2.bias (torch.Size([1024])) -> up_blocks.2.resnets.0.norm2.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.conv2.weight (torch.Size([1024, 1024, 3, 3])) -> up_blocks.2.resnets.0.conv2.weight (torch.Size([512, 512, 3, 3]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.conv2.bias (torch.Size([1024])) -> up_blocks.2.resnets.0.conv2.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.conv_shortcut.weight (torch.Size([1024, 2048, 1, 1])) -> up_blocks.2.resnets.0.conv_shortcut.weight (torch.Size([512, 1536, 1, 1]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.0.conv_shortcut.bias (torch.Size([1024])) -> up_blocks.2.resnets.0.conv_shortcut.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.norm1.weight (torch.Size([2048])) -> up_blocks.2.resnets.1.norm1.weight (torch.Size([1024]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.norm1.bias (torch.Size([2048])) -> up_blocks.2.resnets.1.norm1.bias (torch.Size([1024]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.conv1.weight (torch.Size([1024, 2048, 3, 3])) -> up_blocks.2.resnets.1.conv1.weight (torch.Size([512, 1024, 3, 3]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.conv1.bias (torch.Size([1024])) -> up_blocks.2.resnets.1.conv1.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.time_emb_proj.weight (torch.Size([1024, 4096])) -> up_blocks.2.resnets.1.time_emb_proj.weight (torch.Size([512, 2048]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.time_emb_proj.bias (torch.Size([1024])) -> up_blocks.2.resnets.1.time_emb_proj.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.norm2.weight (torch.Size([1024])) -> up_blocks.2.resnets.1.norm2.weight (torch.Size([512]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.norm2.bias (torch.Size([1024])) -> up_blocks.2.resnets.1.norm2.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.conv2.weight (torch.Size([1024, 1024, 3, 3])) -> up_blocks.2.resnets.1.conv2.weight (torch.Size([512, 512, 3, 3]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.conv2.bias (torch.Size([1024])) -> up_blocks.2.resnets.1.conv2.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.conv_shortcut.weight (torch.Size([1024, 2048, 1, 1])) -> up_blocks.2.resnets.1.conv_shortcut.weight (torch.Size([512, 1024, 1, 1]))\n", + "✗ Несовпадение размеров: up_blocks.2.resnets.1.conv_shortcut.bias (torch.Size([1024])) -> up_blocks.2.resnets.1.conv_shortcut.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: mid_block.resnets.0.time_emb_proj.weight (torch.Size([1280, 4096])) -> mid_block.resnets.0.time_emb_proj.weight (torch.Size([1280, 2048]))\n", + "✗ Несовпадение размеров: mid_block.resnets.1.time_emb_proj.weight (torch.Size([1280, 4096])) -> mid_block.resnets.1.time_emb_proj.weight (torch.Size([1280, 2048]))\n", + "✗ Несовпадение размеров: conv_norm_out.weight (torch.Size([1024])) -> conv_norm_out.weight (torch.Size([512]))\n", + "✗ Несовпадение размеров: conv_norm_out.bias (torch.Size([1024])) -> conv_norm_out.bias (torch.Size([512]))\n", + "✗ Несовпадение размеров: conv_out.weight (torch.Size([128, 1024, 1, 1])) -> conv_out.weight (torch.Size([128, 512, 3, 3]))\n", + "Статистика переноса: {'перенесено': 352, 'несовпадение_размеров': 60, 'пропущено': 492}\n", + "Неперенесенные ключи в новой модели:\n", + "conv_in.bias\n", + "conv_in.weight\n", + "conv_norm_out.bias\n", + "conv_norm_out.weight\n", + "conv_out.weight\n", + "down_blocks.0.downsamplers.0.conv.bias\n", + "down_blocks.0.downsamplers.0.conv.weight\n", + "down_blocks.0.resnets.0.conv1.bias\n", + "down_blocks.0.resnets.0.conv1.weight\n", + "down_blocks.0.resnets.0.conv2.bias\n", + "down_blocks.0.resnets.0.conv2.weight\n", + "down_blocks.0.resnets.0.norm1.bias\n", + "down_blocks.0.resnets.0.norm1.weight\n", + "down_blocks.0.resnets.0.norm2.bias\n", + "down_blocks.0.resnets.0.norm2.weight\n", + "down_blocks.0.resnets.0.time_emb_proj.bias\n", + "down_blocks.0.resnets.0.time_emb_proj.weight\n", + "down_blocks.0.resnets.1.conv1.bias\n", + "down_blocks.0.resnets.1.conv1.weight\n", + "down_blocks.0.resnets.1.conv2.bias\n", + "down_blocks.0.resnets.1.conv2.weight\n", + "down_blocks.0.resnets.1.norm1.bias\n", + "down_blocks.0.resnets.1.norm1.weight\n", + "down_blocks.0.resnets.1.norm2.bias\n", + "down_blocks.0.resnets.1.norm2.weight\n", + "down_blocks.0.resnets.1.time_emb_proj.bias\n", + "down_blocks.0.resnets.1.time_emb_proj.weight\n", + "down_blocks.1.attentions.1.norm.bias\n", + "down_blocks.1.attentions.1.norm.weight\n", + "down_blocks.1.attentions.1.proj_in.bias\n", + "down_blocks.1.attentions.1.proj_in.weight\n", + "down_blocks.1.attentions.1.proj_out.bias\n", + "down_blocks.1.attentions.1.proj_out.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.2.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.2.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.norm1.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.0.norm1.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.norm2.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.0.norm2.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.0.norm3.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.0.norm3.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.attn1.to_k.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.attn1.to_out.0.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.1.attn1.to_out.0.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.attn1.to_q.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.attn1.to_v.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.attn2.to_k.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.attn2.to_out.0.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.1.attn2.to_out.0.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.attn2.to_q.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.attn2.to_v.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.ff.net.0.proj.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.1.ff.net.0.proj.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.ff.net.2.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.1.ff.net.2.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.norm1.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.1.norm1.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.norm2.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.1.norm2.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.1.norm3.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.1.norm3.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.attn1.to_k.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.attn1.to_out.0.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.2.attn1.to_out.0.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.attn1.to_q.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.attn1.to_v.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.attn2.to_k.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.attn2.to_out.0.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.2.attn2.to_out.0.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.attn2.to_q.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.attn2.to_v.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.ff.net.0.proj.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.2.ff.net.0.proj.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.ff.net.2.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.2.ff.net.2.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.norm1.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.2.norm1.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.norm2.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.2.norm2.weight\n", + "down_blocks.1.attentions.1.transformer_blocks.2.norm3.bias\n", + "down_blocks.1.attentions.1.transformer_blocks.2.norm3.weight\n", + "down_blocks.1.resnets.0.conv1.weight\n", + "down_blocks.1.resnets.0.conv_shortcut.bias\n", + "down_blocks.1.resnets.0.conv_shortcut.weight\n", + "down_blocks.1.resnets.0.norm1.bias\n", + "down_blocks.1.resnets.0.norm1.weight\n", + "down_blocks.1.resnets.0.time_emb_proj.weight\n", + "down_blocks.1.resnets.1.conv1.bias\n", + "down_blocks.1.resnets.1.conv1.weight\n", + "down_blocks.1.resnets.1.conv2.bias\n", + "down_blocks.1.resnets.1.conv2.weight\n", + "down_blocks.1.resnets.1.norm1.bias\n", + "down_blocks.1.resnets.1.norm1.weight\n", + "down_blocks.1.resnets.1.norm2.bias\n", + "down_blocks.1.resnets.1.norm2.weight\n", + "down_blocks.1.resnets.1.time_emb_proj.bias\n", + "down_blocks.1.resnets.1.time_emb_proj.weight\n", + "down_blocks.2.attentions.1.norm.bias\n", + "down_blocks.2.attentions.1.norm.weight\n", + "down_blocks.2.attentions.1.proj_in.bias\n", + "down_blocks.2.attentions.1.proj_in.weight\n", + "down_blocks.2.attentions.1.proj_out.bias\n", + "down_blocks.2.attentions.1.proj_out.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.norm1.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.0.norm1.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.norm2.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.0.norm2.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.0.norm3.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.0.norm3.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.attn1.to_k.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.attn1.to_out.0.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.1.attn1.to_out.0.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.attn1.to_q.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.attn1.to_v.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.attn2.to_k.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.attn2.to_out.0.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.1.attn2.to_out.0.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.attn2.to_q.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.attn2.to_v.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.ff.net.0.proj.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.1.ff.net.0.proj.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.ff.net.2.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.1.ff.net.2.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.norm1.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.1.norm1.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.norm2.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.1.norm2.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.1.norm3.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.1.norm3.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.attn1.to_k.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.attn1.to_out.0.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.2.attn1.to_out.0.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.attn1.to_q.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.attn1.to_v.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.attn2.to_k.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.attn2.to_out.0.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.2.attn2.to_out.0.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.attn2.to_q.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.attn2.to_v.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.ff.net.0.proj.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.2.ff.net.0.proj.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.ff.net.2.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.2.ff.net.2.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.norm1.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.2.norm1.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.norm2.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.2.norm2.weight\n", + "down_blocks.2.attentions.1.transformer_blocks.2.norm3.bias\n", + "down_blocks.2.attentions.1.transformer_blocks.2.norm3.weight\n", + "down_blocks.2.resnets.0.time_emb_proj.weight\n", + "down_blocks.2.resnets.1.conv1.bias\n", + "down_blocks.2.resnets.1.conv1.weight\n", + "down_blocks.2.resnets.1.conv2.bias\n", + "down_blocks.2.resnets.1.conv2.weight\n", + "down_blocks.2.resnets.1.norm1.bias\n", + "down_blocks.2.resnets.1.norm1.weight\n", + "down_blocks.2.resnets.1.norm2.bias\n", + "down_blocks.2.resnets.1.norm2.weight\n", + "down_blocks.2.resnets.1.time_emb_proj.bias\n", + "down_blocks.2.resnets.1.time_emb_proj.weight\n", + "mid_block.attentions.0.transformer_blocks.0.attn1.to_k.weight\n", + "mid_block.attentions.0.transformer_blocks.0.attn1.to_out.0.bias\n", + "mid_block.attentions.0.transformer_blocks.0.attn1.to_out.0.weight\n", + "mid_block.attentions.0.transformer_blocks.0.attn1.to_q.weight\n", + "mid_block.attentions.0.transformer_blocks.0.attn1.to_v.weight\n", + "mid_block.attentions.0.transformer_blocks.0.attn2.to_k.weight\n", + "mid_block.attentions.0.transformer_blocks.0.attn2.to_out.0.bias\n", + "mid_block.attentions.0.transformer_blocks.0.attn2.to_out.0.weight\n", + "mid_block.attentions.0.transformer_blocks.0.attn2.to_q.weight\n", + "mid_block.attentions.0.transformer_blocks.0.attn2.to_v.weight\n", + "mid_block.attentions.0.transformer_blocks.0.ff.net.0.proj.bias\n", + "mid_block.attentions.0.transformer_blocks.0.ff.net.0.proj.weight\n", + "mid_block.attentions.0.transformer_blocks.0.ff.net.2.bias\n", + "mid_block.attentions.0.transformer_blocks.0.ff.net.2.weight\n", + "mid_block.attentions.0.transformer_blocks.0.norm1.bias\n", + "mid_block.attentions.0.transformer_blocks.0.norm1.weight\n", + "mid_block.attentions.0.transformer_blocks.0.norm2.bias\n", + "mid_block.attentions.0.transformer_blocks.0.norm2.weight\n", + "mid_block.attentions.0.transformer_blocks.0.norm3.bias\n", + "mid_block.attentions.0.transformer_blocks.0.norm3.weight\n", + "mid_block.attentions.0.transformer_blocks.1.attn1.to_k.weight\n", + "mid_block.attentions.0.transformer_blocks.1.attn1.to_out.0.bias\n", + "mid_block.attentions.0.transformer_blocks.1.attn1.to_out.0.weight\n", + "mid_block.attentions.0.transformer_blocks.1.attn1.to_q.weight\n", + "mid_block.attentions.0.transformer_blocks.1.attn1.to_v.weight\n", + "mid_block.attentions.0.transformer_blocks.1.attn2.to_k.weight\n", + "mid_block.attentions.0.transformer_blocks.1.attn2.to_out.0.bias\n", + "mid_block.attentions.0.transformer_blocks.1.attn2.to_out.0.weight\n", + "mid_block.attentions.0.transformer_blocks.1.attn2.to_q.weight\n", + "mid_block.attentions.0.transformer_blocks.1.attn2.to_v.weight\n", + "mid_block.attentions.0.transformer_blocks.1.ff.net.0.proj.bias\n", + "mid_block.attentions.0.transformer_blocks.1.ff.net.0.proj.weight\n", + "mid_block.attentions.0.transformer_blocks.1.ff.net.2.bias\n", + "mid_block.attentions.0.transformer_blocks.1.ff.net.2.weight\n", + "mid_block.attentions.0.transformer_blocks.1.norm1.bias\n", + "mid_block.attentions.0.transformer_blocks.1.norm1.weight\n", + "mid_block.attentions.0.transformer_blocks.1.norm2.bias\n", + "mid_block.attentions.0.transformer_blocks.1.norm2.weight\n", + "mid_block.attentions.0.transformer_blocks.1.norm3.bias\n", + "mid_block.attentions.0.transformer_blocks.1.norm3.weight\n", + "mid_block.attentions.0.transformer_blocks.2.attn1.to_k.weight\n", + "mid_block.attentions.0.transformer_blocks.2.attn1.to_out.0.bias\n", + "mid_block.attentions.0.transformer_blocks.2.attn1.to_out.0.weight\n", + "mid_block.attentions.0.transformer_blocks.2.attn1.to_q.weight\n", + "mid_block.attentions.0.transformer_blocks.2.attn1.to_v.weight\n", + "mid_block.attentions.0.transformer_blocks.2.attn2.to_k.weight\n", + "mid_block.attentions.0.transformer_blocks.2.attn2.to_out.0.bias\n", + "mid_block.attentions.0.transformer_blocks.2.attn2.to_out.0.weight\n", + "mid_block.attentions.0.transformer_blocks.2.attn2.to_q.weight\n", + "mid_block.attentions.0.transformer_blocks.2.attn2.to_v.weight\n", + "mid_block.attentions.0.transformer_blocks.2.ff.net.0.proj.bias\n", + "mid_block.attentions.0.transformer_blocks.2.ff.net.0.proj.weight\n", + "mid_block.attentions.0.transformer_blocks.2.ff.net.2.bias\n", + "mid_block.attentions.0.transformer_blocks.2.ff.net.2.weight\n", + "mid_block.attentions.0.transformer_blocks.2.norm1.bias\n", + "mid_block.attentions.0.transformer_blocks.2.norm1.weight\n", + "mid_block.attentions.0.transformer_blocks.2.norm2.bias\n", + "mid_block.attentions.0.transformer_blocks.2.norm2.weight\n", + "mid_block.attentions.0.transformer_blocks.2.norm3.bias\n", + "mid_block.attentions.0.transformer_blocks.2.norm3.weight\n", + "mid_block.resnets.0.time_emb_proj.weight\n", + "mid_block.resnets.1.time_emb_proj.weight\n", + "time_embedding.linear_1.bias\n", + "time_embedding.linear_1.weight\n", + "time_embedding.linear_2.bias\n", + "time_embedding.linear_2.weight\n", + "up_blocks.0.attentions.0.norm.bias\n", + "up_blocks.0.attentions.0.norm.weight\n", + "up_blocks.0.attentions.0.proj_in.bias\n", + "up_blocks.0.attentions.0.proj_in.weight\n", + "up_blocks.0.attentions.0.proj_out.bias\n", + "up_blocks.0.attentions.0.proj_out.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.attn1.to_k.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.attn1.to_out.0.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.0.attn1.to_out.0.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.attn1.to_q.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.attn1.to_v.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.attn2.to_k.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.attn2.to_out.0.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.0.attn2.to_out.0.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.attn2.to_q.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.attn2.to_v.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.ff.net.0.proj.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.0.ff.net.0.proj.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.ff.net.2.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.0.ff.net.2.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.norm1.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.0.norm1.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.norm2.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.0.norm2.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.0.norm3.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.0.norm3.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.attn1.to_k.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.attn1.to_out.0.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.1.attn1.to_out.0.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.attn1.to_q.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.attn1.to_v.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.attn2.to_k.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.attn2.to_out.0.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.1.attn2.to_out.0.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.attn2.to_q.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.attn2.to_v.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.ff.net.0.proj.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.1.ff.net.0.proj.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.ff.net.2.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.1.ff.net.2.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.norm1.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.1.norm1.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.norm2.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.1.norm2.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.1.norm3.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.1.norm3.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.attn1.to_k.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.attn1.to_out.0.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.2.attn1.to_out.0.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.attn1.to_q.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.attn1.to_v.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.attn2.to_k.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.attn2.to_out.0.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.2.attn2.to_out.0.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.attn2.to_q.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.attn2.to_v.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.ff.net.0.proj.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.2.ff.net.0.proj.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.ff.net.2.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.2.ff.net.2.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.norm1.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.2.norm1.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.norm2.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.2.norm2.weight\n", + "up_blocks.0.attentions.0.transformer_blocks.2.norm3.bias\n", + "up_blocks.0.attentions.0.transformer_blocks.2.norm3.weight\n", + "up_blocks.0.attentions.1.norm.bias\n", + "up_blocks.0.attentions.1.norm.weight\n", + "up_blocks.0.attentions.1.proj_in.bias\n", + "up_blocks.0.attentions.1.proj_in.weight\n", + "up_blocks.0.attentions.1.proj_out.bias\n", + "up_blocks.0.attentions.1.proj_out.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.attn1.to_k.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.attn1.to_out.0.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.0.attn1.to_out.0.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.attn1.to_q.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.attn1.to_v.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.attn2.to_k.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.attn2.to_out.0.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.0.attn2.to_out.0.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.attn2.to_q.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.attn2.to_v.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.ff.net.0.proj.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.0.ff.net.0.proj.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.ff.net.2.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.0.ff.net.2.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.norm1.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.0.norm1.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.norm2.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.0.norm2.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.0.norm3.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.0.norm3.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.attn1.to_k.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.attn1.to_out.0.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.1.attn1.to_out.0.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.attn1.to_q.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.attn1.to_v.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.attn2.to_k.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.attn2.to_out.0.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.1.attn2.to_out.0.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.attn2.to_q.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.attn2.to_v.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.ff.net.0.proj.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.1.ff.net.0.proj.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.ff.net.2.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.1.ff.net.2.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.norm1.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.1.norm1.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.norm2.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.1.norm2.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.1.norm3.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.1.norm3.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.attn1.to_k.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.attn1.to_out.0.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.2.attn1.to_out.0.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.attn1.to_q.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.attn1.to_v.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.attn2.to_k.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.attn2.to_out.0.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.2.attn2.to_out.0.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.attn2.to_q.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.attn2.to_v.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.ff.net.0.proj.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.2.ff.net.0.proj.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.ff.net.2.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.2.ff.net.2.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.norm1.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.2.norm1.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.norm2.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.2.norm2.weight\n", + "up_blocks.0.attentions.1.transformer_blocks.2.norm3.bias\n", + "up_blocks.0.attentions.1.transformer_blocks.2.norm3.weight\n", + "up_blocks.0.attentions.2.norm.bias\n", + "up_blocks.0.attentions.2.norm.weight\n", + "up_blocks.0.attentions.2.proj_in.bias\n", + "up_blocks.0.attentions.2.proj_in.weight\n", + "up_blocks.0.attentions.2.proj_out.bias\n", + "up_blocks.0.attentions.2.proj_out.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.attn1.to_k.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.attn1.to_out.0.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.0.attn1.to_out.0.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.attn1.to_q.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.attn1.to_v.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.attn2.to_k.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.attn2.to_out.0.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.0.attn2.to_out.0.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.attn2.to_q.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.attn2.to_v.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.ff.net.0.proj.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.0.ff.net.0.proj.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.ff.net.2.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.0.ff.net.2.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.norm1.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.0.norm1.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.norm2.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.0.norm2.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.0.norm3.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.0.norm3.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.attn1.to_k.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.attn1.to_out.0.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.1.attn1.to_out.0.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.attn1.to_q.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.attn1.to_v.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.attn2.to_k.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.attn2.to_out.0.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.1.attn2.to_out.0.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.attn2.to_q.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.attn2.to_v.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.ff.net.0.proj.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.1.ff.net.0.proj.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.ff.net.2.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.1.ff.net.2.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.norm1.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.1.norm1.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.norm2.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.1.norm2.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.1.norm3.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.1.norm3.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.attn1.to_k.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.attn1.to_out.0.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.2.attn1.to_out.0.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.attn1.to_q.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.attn1.to_v.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.attn2.to_k.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.attn2.to_out.0.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.2.attn2.to_out.0.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.attn2.to_q.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.attn2.to_v.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.ff.net.0.proj.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.2.ff.net.0.proj.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.ff.net.2.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.2.ff.net.2.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.norm1.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.2.norm1.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.norm2.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.2.norm2.weight\n", + "up_blocks.0.attentions.2.transformer_blocks.2.norm3.bias\n", + "up_blocks.0.attentions.2.transformer_blocks.2.norm3.weight\n", + "up_blocks.0.resnets.0.time_emb_proj.weight\n", + "up_blocks.0.resnets.1.conv1.weight\n", + "up_blocks.0.resnets.1.conv_shortcut.weight\n", + "up_blocks.0.resnets.1.norm1.bias\n", + "up_blocks.0.resnets.1.norm1.weight\n", + "up_blocks.0.resnets.1.time_emb_proj.weight\n", + "up_blocks.0.resnets.2.conv1.bias\n", + "up_blocks.0.resnets.2.conv1.weight\n", + "up_blocks.0.resnets.2.conv2.bias\n", + "up_blocks.0.resnets.2.conv2.weight\n", + "up_blocks.0.resnets.2.conv_shortcut.bias\n", + "up_blocks.0.resnets.2.conv_shortcut.weight\n", + "up_blocks.0.resnets.2.norm1.bias\n", + "up_blocks.0.resnets.2.norm1.weight\n", + "up_blocks.0.resnets.2.norm2.bias\n", + "up_blocks.0.resnets.2.norm2.weight\n", + "up_blocks.0.resnets.2.time_emb_proj.bias\n", + "up_blocks.0.resnets.2.time_emb_proj.weight\n", + "up_blocks.1.attentions.2.norm.bias\n", + "up_blocks.1.attentions.2.norm.weight\n", + "up_blocks.1.attentions.2.proj_in.bias\n", + "up_blocks.1.attentions.2.proj_in.weight\n", + "up_blocks.1.attentions.2.proj_out.bias\n", + "up_blocks.1.attentions.2.proj_out.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out.0.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out.0.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out.0.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out.0.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.0.proj.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.0.proj.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.2.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.2.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.norm1.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.0.norm1.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.norm2.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.0.norm2.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.0.norm3.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.0.norm3.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.attn1.to_k.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.attn1.to_out.0.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.1.attn1.to_out.0.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.attn1.to_q.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.attn1.to_v.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.attn2.to_k.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.attn2.to_out.0.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.1.attn2.to_out.0.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.attn2.to_q.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.attn2.to_v.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.ff.net.0.proj.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.1.ff.net.0.proj.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.ff.net.2.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.1.ff.net.2.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.norm1.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.1.norm1.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.norm2.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.1.norm2.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.1.norm3.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.1.norm3.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.attn1.to_k.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.attn1.to_out.0.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.2.attn1.to_out.0.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.attn1.to_q.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.attn1.to_v.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.attn2.to_k.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.attn2.to_out.0.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.2.attn2.to_out.0.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.attn2.to_q.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.attn2.to_v.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.ff.net.0.proj.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.2.ff.net.0.proj.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.ff.net.2.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.2.ff.net.2.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.norm1.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.2.norm1.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.norm2.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.2.norm2.weight\n", + "up_blocks.1.attentions.2.transformer_blocks.2.norm3.bias\n", + "up_blocks.1.attentions.2.transformer_blocks.2.norm3.weight\n", + "up_blocks.1.resnets.0.time_emb_proj.weight\n", + "up_blocks.1.resnets.1.time_emb_proj.weight\n", + "up_blocks.1.resnets.2.conv1.bias\n", + "up_blocks.1.resnets.2.conv1.weight\n", + "up_blocks.1.resnets.2.conv2.bias\n", + "up_blocks.1.resnets.2.conv2.weight\n", + "up_blocks.1.resnets.2.conv_shortcut.bias\n", + "up_blocks.1.resnets.2.conv_shortcut.weight\n", + "up_blocks.1.resnets.2.norm1.bias\n", + "up_blocks.1.resnets.2.norm1.weight\n", + "up_blocks.1.resnets.2.norm2.bias\n", + "up_blocks.1.resnets.2.norm2.weight\n", + "up_blocks.1.resnets.2.time_emb_proj.bias\n", + "up_blocks.1.resnets.2.time_emb_proj.weight\n", + "up_blocks.2.resnets.0.conv1.bias\n", + "up_blocks.2.resnets.0.conv1.weight\n", + "up_blocks.2.resnets.0.conv2.bias\n", + "up_blocks.2.resnets.0.conv2.weight\n", + "up_blocks.2.resnets.0.conv_shortcut.bias\n", + "up_blocks.2.resnets.0.conv_shortcut.weight\n", + "up_blocks.2.resnets.0.norm1.bias\n", + "up_blocks.2.resnets.0.norm1.weight\n", + "up_blocks.2.resnets.0.norm2.bias\n", + "up_blocks.2.resnets.0.norm2.weight\n", + "up_blocks.2.resnets.0.time_emb_proj.bias\n", + "up_blocks.2.resnets.0.time_emb_proj.weight\n", + "up_blocks.2.resnets.1.conv1.bias\n", + "up_blocks.2.resnets.1.conv1.weight\n", + "up_blocks.2.resnets.1.conv2.bias\n", + "up_blocks.2.resnets.1.conv2.weight\n", + "up_blocks.2.resnets.1.conv_shortcut.bias\n", + "up_blocks.2.resnets.1.conv_shortcut.weight\n", + "up_blocks.2.resnets.1.norm1.bias\n", + "up_blocks.2.resnets.1.norm1.weight\n", + "up_blocks.2.resnets.1.norm2.bias\n", + "up_blocks.2.resnets.1.norm2.weight\n", + "up_blocks.2.resnets.1.time_emb_proj.bias\n", + "up_blocks.2.resnets.1.time_emb_proj.weight\n", + "up_blocks.2.resnets.2.conv1.bias\n", + "up_blocks.2.resnets.2.conv1.weight\n", + "up_blocks.2.resnets.2.conv2.bias\n", + "up_blocks.2.resnets.2.conv2.weight\n", + "up_blocks.2.resnets.2.conv_shortcut.bias\n", + "up_blocks.2.resnets.2.conv_shortcut.weight\n", + "up_blocks.2.resnets.2.norm1.bias\n", + "up_blocks.2.resnets.2.norm1.weight\n", + "up_blocks.2.resnets.2.norm2.bias\n", + "up_blocks.2.resnets.2.norm2.weight\n", + "up_blocks.2.resnets.2.time_emb_proj.bias\n", + "up_blocks.2.resnets.2.time_emb_proj.weight\n", + "UNet2DConditionModel(\n", + " (conv_in): Conv2d(128, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_proj): Timesteps()\n", + " (time_embedding): TimestepEmbedding(\n", + " (linear_1): Linear(in_features=512, out_features=2048, bias=True)\n", + " (act): SiLU()\n", + " (linear_2): Linear(in_features=2048, out_features=2048, bias=True)\n", + " )\n", + " (down_blocks): ModuleList(\n", + " (0): DownBlock2D(\n", + " (resnets): ModuleList(\n", + " (0-1): 2 x ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 512, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=512, bias=True)\n", + " (norm2): GroupNorm(32, 512, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " )\n", + " )\n", + " (downsamplers): ModuleList(\n", + " (0): Downsample2D(\n", + " (conv): Conv2d(512, 512, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1))\n", + " )\n", + " )\n", + " )\n", + " (1): CrossAttnDownBlock2D(\n", + " (attentions): ModuleList(\n", + " (0-1): 2 x Transformer2DModel(\n", + " (norm): GroupNorm(32, 1024, eps=1e-06, affine=True)\n", + " (proj_in): Conv2d(1024, 1024, kernel_size=(1, 1), stride=(1, 1))\n", + " (transformer_blocks): ModuleList(\n", + " (0-2): 3 x BasicTransformerBlock(\n", + " (norm1): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)\n", + " (attn1): Attention(\n", + " (to_q): Linear(in_features=1024, out_features=1024, bias=False)\n", + " (to_k): Linear(in_features=1024, out_features=1024, bias=False)\n", + " (to_v): Linear(in_features=1024, out_features=1024, bias=False)\n", + " (to_out): ModuleList(\n", + " (0): Linear(in_features=1024, out_features=1024, bias=True)\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " )\n", + " )\n", + " (norm2): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)\n", + " (attn2): Attention(\n", + " (to_q): Linear(in_features=1024, out_features=1024, bias=False)\n", + " (to_k): Linear(in_features=768, out_features=1024, bias=False)\n", + " (to_v): Linear(in_features=768, out_features=1024, bias=False)\n", + " (to_out): ModuleList(\n", + " (0): Linear(in_features=1024, out_features=1024, bias=True)\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " )\n", + " )\n", + " (norm3): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)\n", + " (ff): FeedForward(\n", + " (net): ModuleList(\n", + " (0): GEGLU(\n", + " (proj): Linear(in_features=1024, out_features=8192, bias=True)\n", + " )\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " (2): Linear(in_features=4096, out_features=1024, bias=True)\n", + " )\n", + " )\n", + " )\n", + " )\n", + " (proj_out): Conv2d(1024, 1024, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " )\n", + " (resnets): ModuleList(\n", + " (0): ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 512, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(512, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=1024, bias=True)\n", + " (norm2): GroupNorm(32, 1024, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " (conv_shortcut): Conv2d(512, 1024, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " (1): ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 1024, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=1024, bias=True)\n", + " (norm2): GroupNorm(32, 1024, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " )\n", + " )\n", + " (downsamplers): ModuleList(\n", + " (0): Downsample2D(\n", + " (conv): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1))\n", + " )\n", + " )\n", + " )\n", + " (2): CrossAttnDownBlock2D(\n", + " (attentions): ModuleList(\n", + " (0-1): 2 x Transformer2DModel(\n", + " (norm): GroupNorm(32, 1280, eps=1e-06, affine=True)\n", + " (proj_in): Conv2d(1280, 1280, kernel_size=(1, 1), stride=(1, 1))\n", + " (transformer_blocks): ModuleList(\n", + " (0-2): 3 x BasicTransformerBlock(\n", + " (norm1): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)\n", + " (attn1): Attention(\n", + " (to_q): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_k): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_v): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_out): ModuleList(\n", + " (0): Linear(in_features=1280, out_features=1280, bias=True)\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " )\n", + " )\n", + " (norm2): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)\n", + " (attn2): Attention(\n", + " (to_q): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_k): Linear(in_features=768, out_features=1280, bias=False)\n", + " (to_v): Linear(in_features=768, out_features=1280, bias=False)\n", + " (to_out): ModuleList(\n", + " (0): Linear(in_features=1280, out_features=1280, bias=True)\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " )\n", + " )\n", + " (norm3): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)\n", + " (ff): FeedForward(\n", + " (net): ModuleList(\n", + " (0): GEGLU(\n", + " (proj): Linear(in_features=1280, out_features=10240, bias=True)\n", + " )\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " (2): Linear(in_features=5120, out_features=1280, bias=True)\n", + " )\n", + " )\n", + " )\n", + " )\n", + " (proj_out): Conv2d(1280, 1280, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " )\n", + " (resnets): ModuleList(\n", + " (0): ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 1024, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(1024, 1280, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=1280, bias=True)\n", + " (norm2): GroupNorm(32, 1280, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(1280, 1280, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " (conv_shortcut): Conv2d(1024, 1280, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " (1): ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 1280, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(1280, 1280, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=1280, bias=True)\n", + " (norm2): GroupNorm(32, 1280, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(1280, 1280, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " )\n", + " )\n", + " )\n", + " )\n", + " (up_blocks): ModuleList(\n", + " (0): CrossAttnUpBlock2D(\n", + " (attentions): ModuleList(\n", + " (0-2): 3 x Transformer2DModel(\n", + " (norm): GroupNorm(32, 1280, eps=1e-06, affine=True)\n", + " (proj_in): Conv2d(1280, 1280, kernel_size=(1, 1), stride=(1, 1))\n", + " (transformer_blocks): ModuleList(\n", + " (0-2): 3 x BasicTransformerBlock(\n", + " (norm1): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)\n", + " (attn1): Attention(\n", + " (to_q): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_k): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_v): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_out): ModuleList(\n", + " (0): Linear(in_features=1280, out_features=1280, bias=True)\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " )\n", + " )\n", + " (norm2): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)\n", + " (attn2): Attention(\n", + " (to_q): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_k): Linear(in_features=768, out_features=1280, bias=False)\n", + " (to_v): Linear(in_features=768, out_features=1280, bias=False)\n", + " (to_out): ModuleList(\n", + " (0): Linear(in_features=1280, out_features=1280, bias=True)\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " )\n", + " )\n", + " (norm3): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)\n", + " (ff): FeedForward(\n", + " (net): ModuleList(\n", + " (0): GEGLU(\n", + " (proj): Linear(in_features=1280, out_features=10240, bias=True)\n", + " )\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " (2): Linear(in_features=5120, out_features=1280, bias=True)\n", + " )\n", + " )\n", + " )\n", + " )\n", + " (proj_out): Conv2d(1280, 1280, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " )\n", + " (resnets): ModuleList(\n", + " (0-1): 2 x ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 2560, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(2560, 1280, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=1280, bias=True)\n", + " (norm2): GroupNorm(32, 1280, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(1280, 1280, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " (conv_shortcut): Conv2d(2560, 1280, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " (2): ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 2304, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(2304, 1280, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=1280, bias=True)\n", + " (norm2): GroupNorm(32, 1280, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(1280, 1280, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " (conv_shortcut): Conv2d(2304, 1280, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " )\n", + " (upsamplers): ModuleList(\n", + " (0): Upsample2D(\n", + " (conv): Conv2d(1280, 1280, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " )\n", + " )\n", + " )\n", + " (1): CrossAttnUpBlock2D(\n", + " (attentions): ModuleList(\n", + " (0-2): 3 x Transformer2DModel(\n", + " (norm): GroupNorm(32, 1024, eps=1e-06, affine=True)\n", + " (proj_in): Conv2d(1024, 1024, kernel_size=(1, 1), stride=(1, 1))\n", + " (transformer_blocks): ModuleList(\n", + " (0-2): 3 x BasicTransformerBlock(\n", + " (norm1): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)\n", + " (attn1): Attention(\n", + " (to_q): Linear(in_features=1024, out_features=1024, bias=False)\n", + " (to_k): Linear(in_features=1024, out_features=1024, bias=False)\n", + " (to_v): Linear(in_features=1024, out_features=1024, bias=False)\n", + " (to_out): ModuleList(\n", + " (0): Linear(in_features=1024, out_features=1024, bias=True)\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " )\n", + " )\n", + " (norm2): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)\n", + " (attn2): Attention(\n", + " (to_q): Linear(in_features=1024, out_features=1024, bias=False)\n", + " (to_k): Linear(in_features=768, out_features=1024, bias=False)\n", + " (to_v): Linear(in_features=768, out_features=1024, bias=False)\n", + " (to_out): ModuleList(\n", + " (0): Linear(in_features=1024, out_features=1024, bias=True)\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " )\n", + " )\n", + " (norm3): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)\n", + " (ff): FeedForward(\n", + " (net): ModuleList(\n", + " (0): GEGLU(\n", + " (proj): Linear(in_features=1024, out_features=8192, bias=True)\n", + " )\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " (2): Linear(in_features=4096, out_features=1024, bias=True)\n", + " )\n", + " )\n", + " )\n", + " )\n", + " (proj_out): Conv2d(1024, 1024, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " )\n", + " (resnets): ModuleList(\n", + " (0): ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 2304, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(2304, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=1024, bias=True)\n", + " (norm2): GroupNorm(32, 1024, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " (conv_shortcut): Conv2d(2304, 1024, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " (1): ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 2048, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(2048, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=1024, bias=True)\n", + " (norm2): GroupNorm(32, 1024, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " (conv_shortcut): Conv2d(2048, 1024, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " (2): ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 1536, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(1536, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=1024, bias=True)\n", + " (norm2): GroupNorm(32, 1024, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " (conv_shortcut): Conv2d(1536, 1024, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " )\n", + " (upsamplers): ModuleList(\n", + " (0): Upsample2D(\n", + " (conv): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " )\n", + " )\n", + " )\n", + " (2): UpBlock2D(\n", + " (resnets): ModuleList(\n", + " (0): ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 1536, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(1536, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=512, bias=True)\n", + " (norm2): GroupNorm(32, 512, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " (conv_shortcut): Conv2d(1536, 512, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " (1-2): 2 x ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 1024, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(1024, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=512, bias=True)\n", + " (norm2): GroupNorm(32, 512, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " (conv_shortcut): Conv2d(1024, 512, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " )\n", + " )\n", + " )\n", + " (mid_block): UNetMidBlock2DCrossAttn(\n", + " (attentions): ModuleList(\n", + " (0): Transformer2DModel(\n", + " (norm): GroupNorm(32, 1280, eps=1e-06, affine=True)\n", + " (proj_in): Conv2d(1280, 1280, kernel_size=(1, 1), stride=(1, 1))\n", + " (transformer_blocks): ModuleList(\n", + " (0-2): 3 x BasicTransformerBlock(\n", + " (norm1): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)\n", + " (attn1): Attention(\n", + " (to_q): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_k): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_v): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_out): ModuleList(\n", + " (0): Linear(in_features=1280, out_features=1280, bias=True)\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " )\n", + " )\n", + " (norm2): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)\n", + " (attn2): Attention(\n", + " (to_q): Linear(in_features=1280, out_features=1280, bias=False)\n", + " (to_k): Linear(in_features=768, out_features=1280, bias=False)\n", + " (to_v): Linear(in_features=768, out_features=1280, bias=False)\n", + " (to_out): ModuleList(\n", + " (0): Linear(in_features=1280, out_features=1280, bias=True)\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " )\n", + " )\n", + " (norm3): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)\n", + " (ff): FeedForward(\n", + " (net): ModuleList(\n", + " (0): GEGLU(\n", + " (proj): Linear(in_features=1280, out_features=10240, bias=True)\n", + " )\n", + " (1): Dropout(p=0.0, inplace=False)\n", + " (2): Linear(in_features=5120, out_features=1280, bias=True)\n", + " )\n", + " )\n", + " )\n", + " )\n", + " (proj_out): Conv2d(1280, 1280, kernel_size=(1, 1), stride=(1, 1))\n", + " )\n", + " )\n", + " (resnets): ModuleList(\n", + " (0-1): 2 x ResnetBlock2D(\n", + " (norm1): GroupNorm(32, 1280, eps=1e-05, affine=True)\n", + " (conv1): Conv2d(1280, 1280, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (time_emb_proj): Linear(in_features=2048, out_features=1280, bias=True)\n", + " (norm2): GroupNorm(32, 1280, eps=1e-05, affine=True)\n", + " (dropout): Dropout(p=0.0, inplace=False)\n", + " (conv2): Conv2d(1280, 1280, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + " (nonlinearity): SiLU()\n", + " )\n", + " )\n", + " )\n", + " (conv_norm_out): GroupNorm(32, 512, eps=1e-05, affine=True)\n", + " (conv_act): SiLU()\n", + " (conv_out): Conv2d(512, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))\n", + ")\n" + ] + } + ], + "source": [ + "import torch\n", + "from diffusers import UNet2DConditionModel\n", + "from tqdm import tqdm\n", + "\n", + "def log(message):\n", + " print(message)\n", + "\n", + "def main():\n", + " checkpoint_path_old = \"unet_stas\"\n", + " checkpoint_path_new = \"tmp\"\n", + " device = \"cuda\"\n", + " dtype = torch.float16\n", + "\n", + " # Загрузка моделей\n", + " old_unet = UNet2DConditionModel.from_pretrained(checkpoint_path_old).to(device, dtype=dtype)\n", + " new_unet = UNet2DConditionModel.from_pretrained(checkpoint_path_new).to(device, dtype=dtype)\n", + "\n", + " old_state_dict = old_unet.state_dict()\n", + " new_state_dict = new_unet.state_dict()\n", + "\n", + " transferred_state_dict = {}\n", + " transfer_stats = {\n", + " \"перенесено\": 0,\n", + " \"несовпадение_размеров\": 0,\n", + " \"пропущено\": 0\n", + " }\n", + "\n", + " transferred_keys = set()\n", + "\n", + " # Обрабатываем каждый ключ старой модели\n", + " for old_key in tqdm(old_state_dict.keys(), desc=\"Перенос весов\"):\n", + " new_key = old_key\n", + "\n", + " # Проверяем, существует ли ключ в новой модели\n", + " if new_key in new_state_dict:\n", + " # Проверяем совместимость размеров\n", + " if old_state_dict[old_key].shape == new_state_dict[new_key].shape:\n", + " transferred_state_dict[new_key] = old_state_dict[old_key].clone()\n", + " transferred_keys.add(new_key)\n", + " transfer_stats[\"перенесено\"] += 1\n", + " #log(f\"✓ Перенос: {old_key} -> {new_key}, форма: {old_state_dict[old_key].shape}\")\n", + " else:\n", + " log(f\"✗ Несовпадение размеров: {old_key} ({old_state_dict[old_key].shape}) -> {new_key} ({new_state_dict[new_key].shape})\")\n", + " transfer_stats[\"несовпадение_размеров\"] += 1\n", + " else:\n", + " log(f\"? Ключ не найден в новой модели: {old_key} -> {old_state_dict[old_key].shape}\")\n", + " transfer_stats[\"пропущено\"] += 1\n", + "\n", + " # Обновляем состояние новой модели перенесенными весами\n", + " new_state_dict.update(transferred_state_dict)\n", + " new_unet.load_state_dict(new_state_dict)\n", + " new_unet.save_pretrained(\"unet\")\n", + "\n", + " # Получаем список неперенесенных ключей\n", + " non_transferred_keys = sorted(set(new_state_dict.keys()) - transferred_keys)\n", + "\n", + " print(\"Статистика переноса:\", transfer_stats)\n", + " print(\"Неперенесенные ключи в новой модели:\")\n", + " for key in non_transferred_keys:\n", + " print(key)\n", + "\n", + " print(new_unet)\n", + "\n", + "if __name__ == \"__main__\":\n", + " main()\n", + "# Статистика переноса: {'перенесено': 686, 'несовпадение_размеров': 0, 'пропущено': 0}" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "f2438e3d-4b83-4b3f-8e78-53cbcc35f6e4", + "metadata": {}, + "outputs": [], + "source": [] + } + ], + "metadata": { + "kernelspec": { + "display_name": "Python 3 (ipykernel)", + "language": "python", + "name": "python3" + }, + "language_info": { + "codemirror_mode": { + "name": "ipython", + "version": 3 + }, + "file_extension": ".py", + "mimetype": "text/x-python", + "name": "python", + "nbconvert_exporter": "python", + "pygments_lexer": "ipython3", + "version": "3.11.6" + } + }, + "nbformat": 4, + "nbformat_minor": 5 +}