| |
| |
|
|
| import math |
| from typing import Callable, List, Optional, Tuple, Union |
| from collections import defaultdict |
| import os |
| import shutil |
| import cv2 |
| import numpy as np |
| import einops |
|
|
| import torch |
| import torch.nn as nn |
| import torch.nn.functional as F |
|
|
| from .mit48px_ctc import AvgMeter, chunks, TextBlock |
|
|
|
|
| def fixed_pos_embedding(x): |
| seq_len, dim = x.shape |
| inv_freq = 1.0 / (10000 ** (torch.arange(0, dim) / dim)) |
| sinusoid_inp = ( |
| torch.einsum("i , j -> i j", torch.arange(0, seq_len, dtype=torch.float), inv_freq).to(x) |
| ) |
| return torch.sin(sinusoid_inp), torch.cos(sinusoid_inp) |
|
|
| def rotate_every_two(x): |
| x1 = x[:, :, ::2] |
| x2 = x[:, :, 1::2] |
| x = torch.stack((-x2, x1), dim=-1) |
| return x.flatten(-2) |
|
|
| def duplicate_interleave(m): |
| """ |
| A simple version of `torch.repeat_interleave` for duplicating a matrix while interleaving the copy. |
| """ |
| dim0 = m.shape[0] |
| m = m.view(-1, 1) |
| m = m.repeat(1, 2) |
| m = m.view(dim0, -1) |
| return m |
|
|
| def apply_rotary_pos_emb(x, sin, cos, scale=1): |
| sin, cos = map(lambda t: duplicate_interleave(t * scale), (sin, cos)) |
| |
| return (x * cos) + (rotate_every_two(x) * sin) |
|
|
| def apply_rotary_pos_emb2d(x, sin, cos, scale=1): |
| breakpoint() |
| sin, cos = map(lambda t: duplicate_interleave(t * scale), (sin, cos)) |
| |
| return (x * cos) + (rotate_every_two(x) * sin) |
|
|
| class XPOS(nn.Module): |
| def __init__( |
| self, head_dim, scale_base=512 |
| ): |
| super().__init__() |
| self.head_dim = head_dim |
| self.scale_base = scale_base |
| self.register_buffer( |
| "scale", (torch.arange(0, head_dim, 2) + 0.4 * head_dim) / (1.4 * head_dim) |
| ) |
|
|
| def forward(self, x, offset=0, downscale=False): |
| length = x.shape[1] |
| min_pos = -(length + offset) // 2 |
| max_pos = length + offset + min_pos |
| scale = self.scale ** torch.arange(min_pos, max_pos, 1).to(self.scale).div(self.scale_base)[:, None] |
| sin, cos = fixed_pos_embedding(scale) |
|
|
| if scale.shape[0] > length: |
| scale = scale[-length:] |
| sin = sin[-length:] |
| cos = cos[-length:] |
| |
| if downscale: |
| scale = 1 / scale |
|
|
| x = apply_rotary_pos_emb(x, sin, cos, scale) |
| return x |
|
|
|
|
| class XPOS2D(nn.Module): |
| def __init__( |
| self, head_dim, scale_base=512 |
| ): |
| super().__init__() |
| self.xpos = XPOS(head_dim // 2, scale_base) |
|
|
| def forward(self, x: torch.Tensor, offset_x = 0, offset_y = 0, downscale=False): |
| """ |
| x: N, H, W, C |
| """ |
| N, H, W, C = x.shape |
| C = C // 2 |
| [dir_x, dir_y] = x.chunk(2, dim = 3) |
| dir_x = einops.rearrange(dir_x, 'N H W C -> (N H) W C', N = N, H = H, W = W, C = C) |
| dir_y = einops.rearrange(dir_y, 'N H W C -> (N W) H C', N = N, H = H, W = W, C = C) |
| dir_x = self.xpos(dir_x, offset = offset_x, downscale = downscale) |
| dir_y = self.xpos(dir_y, offset = offset_y, downscale = downscale) |
| dir_x = einops.rearrange(dir_x, '(N H) W C -> N H W C', N = N, H = H, W = W, C = C) |
| dir_y = einops.rearrange(dir_y, '(N W) H C -> N H W C', N = N, H = H, W = W, C = C) |
| return torch.cat([dir_x, dir_y], dim = 3) |
|
|
|
|
| |
|
|
| class Model48pxOCR: |
| _MODEL_MAPPING = { |
| 'model': { |
| 'url': 'https://huggingface.co/zyddnys/manga-image-translator/resolve/main/ocr_ar_48px.ckpt', |
| 'hash': '29daa46d080818bb4ab239a518a88338cbccff8f901bef8c9db191a7cb97671d', |
| }, |
| 'dict': { |
| 'url': 'https://huggingface.co/zyddnys/manga-image-translator/resolve/main/alphabet-all-v7.txt', |
| 'hash': 'f5722368146aa0fbcc9f4726866e4efc3203318ebb66c811d8cbbe915576538a', |
| }, |
| } |
|
|
| def __init__(self, model_path: str, device='cpu', *args, **kwargs): |
|
|
| super().__init__(*args, **kwargs) |
|
|
| self.device = device |
| self.text_height = 48 |
| self.maxwidth = 8100 |
|
|
| with open('data/alphabet-all-v7.txt', 'r', encoding = 'utf-8') as fp: |
| dictionary = [s[:-1] for s in fp.readlines()] |
| self.model = OCR(dictionary, 768) |
| sd = torch.load('data/models/ocr_ar_48px.ckpt', map_location='cpu') |
| self.model.load_state_dict(sd) |
| self.model.eval() |
| if self.device != 'cpu' : |
| self.model = self.model.to(self.device) |
|
|
| def to(self, device: str) -> None: |
| self.model.to(device) |
| self.device = device |
| |
| def __call__(self, textblk_lst: List[TextBlock], regions: List[np.ndarray], textblk_lst_indices: List, chunk_size = 16) -> None: |
| perm = range(len(regions)) |
| chunck_idx = 0 |
| for indices in chunks(perm, chunk_size): |
| N = len(indices) |
| widths = [regions[i].shape[1] for i in indices] |
| max_width = 4 * (max(widths) + 7) // 4 |
| region = np.zeros((N, self.text_height, max_width, 3), dtype = np.uint8) |
| for i, idx in enumerate(indices): |
| W = regions[idx].shape[1] |
| |
| region_data = regions[idx] |
| region[i, :, : W, :]=region_data |
|
|
| image_tensor = (torch.from_numpy(region).float() - 127.5) / 127.5 |
| image_tensor = einops.rearrange(image_tensor, 'N H W C -> N C H W') |
| |
| if self.device != 'cpu': |
| image_tensor = image_tensor.to(self.device) |
|
|
| with torch.no_grad(): |
| ret = self.model.infer_beam_batch_tensor(image_tensor, widths, beams_k = 5, max_seq_length = 255) |
| for i, (pred_chars_index, prob, fg_pred, bg_pred, fg_ind_pred, bg_ind_pred) in enumerate(ret): |
| if prob < 0.2: |
| continue |
| has_fg = (fg_ind_pred[:, 1] > fg_ind_pred[:, 0]) |
| has_bg = (bg_ind_pred[:, 1] > bg_ind_pred[:, 0]) |
| seq = [] |
| fr = AvgMeter() |
| fg = AvgMeter() |
| fb = AvgMeter() |
| br = AvgMeter() |
| bg = AvgMeter() |
| bb = AvgMeter() |
| for chid, c_fg, c_bg, h_fg, h_bg in zip(pred_chars_index, fg_pred, bg_pred, has_fg, has_bg) : |
| ch = self.model.dictionary[chid] |
| if ch == '<S>': |
| continue |
| if ch == '</S>': |
| break |
| if ch == '<SP>': |
| ch = ' ' |
| seq.append(ch) |
| if h_fg.item() : |
| fr(int(c_fg[0] * 255)) |
| fg(int(c_fg[1] * 255)) |
| fb(int(c_fg[2] * 255)) |
| if h_bg.item() : |
| br(int(c_bg[0] * 255)) |
| bg(int(c_bg[1] * 255)) |
| bb(int(c_bg[2] * 255)) |
| else : |
| br(int(c_fg[0] * 255)) |
| bg(int(c_fg[1] * 255)) |
| bb(int(c_fg[2] * 255)) |
| txt = ''.join(seq) |
| fr = min(max(int(fr()), 0), 255) |
| fg = min(max(int(fg()), 0), 255) |
| fb = min(max(int(fb()), 0), 255) |
| br = min(max(int(br()), 0), 255) |
| bg = min(max(int(bg()), 0), 255) |
| bb = min(max(int(bb()), 0), 255) |
| |
| |
| cur_region = textblk_lst[textblk_lst_indices[i+chunck_idx]] |
| cur_region.text.append(txt) |
| cur_region.update_font_colors(np.array([fr, fg, fb]), np.array([br, bg, bb])) |
|
|
| chunck_idx += N |
|
|
|
|
|
|
| class ConvNeXtBlock(nn.Module): |
| r""" ConvNeXt Block. There are two equivalent implementations: |
| (1) DwConv -> LayerNorm (channels_first) -> 1x1 Conv -> GELU -> 1x1 Conv; all in (N, C, H, W) |
| (2) DwConv -> Permute to (N, H, W, C); LayerNorm (channels_last) -> Linear -> GELU -> Linear; Permute back |
| We use (2) as we find it slightly faster in PyTorch |
| |
| Args: |
| dim (int): Number of input channels. |
| drop_path (float): Stochastic depth rate. Default: 0.0 |
| layer_scale_init_value (float): Init value for Layer Scale. Default: 1e-6. |
| """ |
| def __init__(self, dim, layer_scale_init_value=1e-6, ks = 7, padding = 3): |
| super().__init__() |
| self.dwconv = nn.Conv2d(dim, dim, kernel_size=ks, padding=padding, groups=dim) |
| self.norm = nn.BatchNorm2d(dim, eps=1e-6) |
| self.pwconv1 = nn.Conv2d(dim, 4 * dim, 1, 1, 0) |
| self.act = nn.GELU() |
| self.pwconv2 = nn.Conv2d(4 * dim, dim, 1, 1, 0) |
| self.gamma = nn.Parameter(layer_scale_init_value * torch.ones(1, dim, 1, 1), |
| requires_grad=True) if layer_scale_init_value > 0 else None |
|
|
| def forward(self, x): |
| input = x |
| x = self.dwconv(x) |
| x = self.norm(x) |
| x = self.pwconv1(x) |
| x = self.act(x) |
| x = self.pwconv2(x) |
| if self.gamma is not None: |
| x = self.gamma * x |
|
|
| x = input + x |
| return x |
|
|
| class ConvNext_FeatureExtractor(nn.Module) : |
| def __init__(self, img_height = 48, in_dim = 3, dim = 512, n_layers = 12) -> None: |
| super().__init__() |
| base = dim // 8 |
| self.stem = nn.Sequential( |
| nn.Conv2d(in_dim, base, kernel_size = 7, stride = 1, padding = 3), |
| nn.BatchNorm2d(base), |
| nn.ReLU(), |
| nn.Conv2d(base, base * 2, kernel_size = 2, stride = 2, padding = 0), |
| nn.BatchNorm2d(base * 2), |
| nn.ReLU(), |
| nn.Conv2d(base * 2, base * 2, kernel_size = 3, stride = 1, padding = 1), |
| nn.BatchNorm2d(base * 2), |
| nn.ReLU(), |
| ) |
| self.block1 = self.make_layers(base * 2, 4) |
| self.down1 = nn.Sequential( |
| nn.Conv2d(base * 2, base * 4, kernel_size = 2, stride = 2, padding = 0), |
| nn.BatchNorm2d(base * 4), |
| nn.ReLU(), |
| ) |
| self.block2 = self.make_layers(base * 4, 12) |
| self.down2 = nn.Sequential( |
| nn.Conv2d(base * 4, base * 8, kernel_size = (2, 1), stride = (2, 1), padding = (0, 0)), |
| nn.BatchNorm2d(base * 8), |
| nn.ReLU(), |
| ) |
| self.block3 = self.make_layers(base * 8, 10, ks = 5, padding = 2) |
| self.down3 = nn.Sequential( |
| nn.Conv2d(base * 8, base * 8, kernel_size = (2, 1), stride = (2, 1), padding = (0, 0)), |
| nn.BatchNorm2d(base * 8), |
| nn.ReLU(), |
| ) |
| self.block4 = self.make_layers(base * 8, 8, ks = 3, padding = 1) |
| self.down4 = nn.Sequential( |
| nn.Conv2d(base * 8, base * 8, kernel_size = (3, 1), stride = (1, 1), padding = (0, 0)), |
| nn.BatchNorm2d(base * 8), |
| nn.ReLU(), |
| ) |
|
|
| def make_layers(self, dim, n, ks = 7, padding = 3) : |
| layers = [] |
| for i in range(n) : |
| layers.append(ConvNeXtBlock(dim, ks = ks, padding = padding)) |
| return nn.Sequential(*layers) |
| |
| def forward(self, x) : |
| x = self.stem(x) |
| |
| x = self.block1(x) |
| x = self.down1(x) |
| |
| x = self.block2(x) |
| x = self.down2(x) |
| |
| x = self.block3(x) |
| x = self.down3(x) |
| |
| x = self.block4(x) |
| x = self.down4(x) |
| return x |
|
|
| def transformer_encoder_forward( |
| self, |
| src: torch.Tensor, |
| src_mask: Optional[torch.Tensor] = None, |
| src_key_padding_mask: Optional[torch.Tensor] = None, |
| is_causal: bool = False) -> torch.Tensor: |
| x = src |
| if self.norm_first: |
| x = x + self._sa_block(self.norm1(x), src_mask, src_key_padding_mask) |
| x = x + self._ff_block(self.norm2(x)) |
| else: |
| x = self.norm1(x + self._sa_block(x, src_mask, src_key_padding_mask)) |
| x = self.norm2(x + self._ff_block(x)) |
|
|
| return x |
|
|
| class XposMultiheadAttention(nn.Module): |
| def __init__( |
| self, |
| embed_dim, |
| num_heads, |
| self_attention=False, |
| encoder_decoder_attention=False, |
| ): |
| super().__init__() |
| self.embed_dim = embed_dim |
| self.num_heads = num_heads |
| self.head_dim = embed_dim // num_heads |
| self.scaling = self.head_dim**-0.5 |
|
|
| self.self_attention = self_attention |
| self.encoder_decoder_attention = encoder_decoder_attention |
| assert self.self_attention ^ self.encoder_decoder_attention |
|
|
| self.k_proj = nn.Linear(embed_dim, embed_dim, bias = True) |
| self.v_proj = nn.Linear(embed_dim, embed_dim, bias = True) |
| self.q_proj = nn.Linear(embed_dim, embed_dim, bias = True) |
| self.out_proj = nn.Linear(embed_dim, embed_dim, bias = True) |
| self.xpos = XPOS(self.head_dim, embed_dim) |
| self.batch_first = True |
| self._qkv_same_embed_dim = True |
|
|
| def reset_parameters(self): |
| nn.init.xavier_uniform_(self.k_proj.weight, gain=1 / math.sqrt(2)) |
| nn.init.xavier_uniform_(self.v_proj.weight, gain=1 / math.sqrt(2)) |
| nn.init.xavier_uniform_(self.q_proj.weight, gain=1 / math.sqrt(2)) |
| nn.init.xavier_uniform_(self.out_proj.weight) |
| nn.init.constant_(self.out_proj.bias, 0.0) |
|
|
| def forward( |
| self, |
| query, |
| key, |
| value, |
| key_padding_mask=None, |
| attn_mask=None, |
| need_weights = False, |
| is_causal = False, |
| k_offset = 0, |
| q_offset = 0 |
| ): |
| assert not is_causal |
| bsz, tgt_len, embed_dim = query.size() |
| src_len = tgt_len |
| assert embed_dim == self.embed_dim, f"query dim {embed_dim} != {self.embed_dim}" |
|
|
| key_bsz, src_len, _ = key.size() |
| assert key_bsz == bsz, f"{query.size(), key.size()}" |
| assert value is not None |
| assert bsz, src_len == value.shape[:2] |
|
|
| q = self.q_proj(query) |
| k = self.k_proj(key) |
| v = self.v_proj(value) |
| q *= self.scaling |
|
|
| q = q.view(bsz, tgt_len, self.num_heads, self.head_dim).transpose(1, 2) |
| k = k.view(bsz, src_len, self.num_heads, self.head_dim).transpose(1, 2) |
| v = v.view(bsz, src_len, self.num_heads, self.head_dim).transpose(1, 2) |
| q = q.reshape(bsz * self.num_heads, tgt_len, self.head_dim) |
| k = k.reshape(bsz * self.num_heads, src_len, self.head_dim) |
| v = v.reshape(bsz * self.num_heads, src_len, self.head_dim) |
|
|
| if self.xpos is not None: |
| k = self.xpos(k, offset=k_offset, downscale=True) |
| q = self.xpos(q, offset=q_offset, downscale=False) |
|
|
| attn_weights = torch.bmm(q, k.transpose(1, 2)) |
|
|
| if attn_mask is not None: |
| attn_weights = torch.nan_to_num(attn_weights) |
| attn_mask = attn_mask.unsqueeze(0) |
| attn_weights += attn_mask |
|
|
| if key_padding_mask is not None: |
| attn_weights = attn_weights.view(bsz, self.num_heads, tgt_len, src_len) |
| attn_weights = attn_weights.masked_fill( |
| key_padding_mask.unsqueeze(1).unsqueeze(2).to(torch.bool), |
| float("-inf"), |
| ) |
| attn_weights = attn_weights.view(bsz * self.num_heads, tgt_len, src_len) |
|
|
| attn_weights = F.softmax(attn_weights, dim=-1, dtype=torch.float32).type_as( |
| attn_weights |
| ) |
| attn = torch.bmm(attn_weights, v) |
| attn = attn.transpose(0, 1).reshape(tgt_len, bsz, embed_dim).transpose(0, 1) |
|
|
| attn = self.out_proj(attn) |
| attn_weights = attn_weights.view( |
| bsz, self.num_heads, tgt_len, src_len |
| ).transpose(1, 0) |
|
|
| if need_weights: |
| return attn, attn_weights |
| else : |
| return attn, None |
|
|
| def generate_square_subsequent_mask(sz): |
| mask = (torch.triu(torch.ones(sz, sz)) == 1).transpose(0, 1) |
| mask = mask.float().masked_fill(mask == 0, float('-inf')).masked_fill(mask == 1, float(0.0)) |
| return mask |
|
|
| class Beam: |
| def __init__(self, char_seq = [], logprobs = []): |
| |
| if isinstance(char_seq, list): |
| self.chars = torch.tensor(char_seq, dtype=torch.long) |
| self.logprobs = torch.tensor(logprobs, dtype=torch.float32) |
| else: |
| self.chars = char_seq.clone() |
| self.logprobs = logprobs.clone() |
|
|
| def avg_logprob(self): |
| return self.logprobs.mean().item() |
|
|
| def sort_key(self): |
| return -self.avg_logprob() |
|
|
| def seq_end(self, end_tok): |
| return self.chars.view(-1)[-1] == end_tok |
|
|
| def extend(self, idx, logprob): |
| return Beam( |
| torch.cat([self.chars, idx.unsqueeze(0)], dim = -1), |
| torch.cat([self.logprobs, logprob.unsqueeze(0)], dim = -1), |
| ) |
|
|
| DECODE_BLOCK_LENGTH = 8 |
|
|
| class Hypothesis: |
| def __init__(self, device, start_tok: int, end_tok: int, padding_tok: int, memory_idx: int, num_layers: int, embd_dim: int): |
| self.device = device |
| self.start_tok = start_tok |
| self.end_tok = end_tok |
| self.padding_tok = padding_tok |
| self.memory_idx = memory_idx |
| self.embd_size = embd_dim |
| self.num_layers = num_layers |
| |
| self.cached_activations = [torch.zeros(1, 0, self.embd_size).to(self.device)] * (num_layers + 1) |
| self.out_idx = torch.LongTensor([start_tok]).to(self.device) |
| self.out_logprobs = torch.FloatTensor([0]).to(self.device) |
| self.length = 0 |
|
|
| def seq_end(self): |
| return self.out_idx.view(-1)[-1] == self.end_tok |
|
|
| def logprob(self): |
| return self.out_logprobs.mean().item() |
|
|
| def sort_key(self): |
| return -self.logprob() |
|
|
| def prob(self): |
| return self.out_logprobs.mean().exp().item() |
|
|
| def __len__(self): |
| return self.length |
|
|
| def extend(self, idx, logprob): |
| ret = Hypothesis(self.device, self.start_tok, self.end_tok, self.padding_tok, self.memory_idx, self.num_layers, self.embd_size) |
| ret.cached_activations = [item.clone() for item in self.cached_activations] |
| ret.length = self.length + 1 |
| ret.out_idx = torch.cat([self.out_idx, torch.LongTensor([idx]).to(self.device)], dim = 0) |
| ret.out_logprobs = torch.cat([self.out_logprobs, torch.FloatTensor([logprob]).to(self.device)], dim = 0) |
| return ret |
|
|
| def output(self): |
| return self.cached_activations[-1] |
|
|
| def next_token_batch( |
| hyps: List[Hypothesis], |
| memory: torch.Tensor, |
| memory_mask: torch.BoolTensor, |
| decoders: nn.ModuleList, |
| embd: nn.Embedding |
| ): |
| layer: nn.TransformerDecoderLayer |
| N = len(hyps) |
| offset = len(hyps[0]) |
|
|
| |
| last_toks = torch.stack([item.out_idx[-1] for item in hyps]) |
| |
| tgt: torch.FloatTensor = embd(last_toks).unsqueeze_(1) |
| |
| |
| memory = torch.stack([memory[idx, :, :] for idx in [item.memory_idx for item in hyps]], dim = 0) |
| for l, layer in enumerate(decoders): |
| |
| |
| combined_activations = torch.cat([item.cached_activations[l] for item in hyps], dim = 0) |
| |
| combined_activations = torch.cat([combined_activations, tgt], dim = 1) |
| for i in range(N): |
| hyps[i].cached_activations[l] = combined_activations[i: i + 1, :, :] |
| |
| tgt = tgt + layer.self_attn(layer.norm1(tgt), layer.norm1(combined_activations), layer.norm1(combined_activations), q_offset = offset)[0] |
| tgt = tgt + layer.multihead_attn(layer.norm2(tgt), memory, memory, key_padding_mask = memory_mask, q_offset = offset)[0] |
| tgt = tgt + layer._ff_block(layer.norm3(tgt)) |
| |
| for i in range(N): |
| hyps[i].cached_activations[len(decoders)] = torch.cat([hyps[i].cached_activations[len(decoders)], tgt[i: i + 1, :, :]], dim = 1) |
| |
| return tgt.squeeze_(1) |
|
|
| class OCR(nn.Module): |
| def __init__(self, dictionary, max_len): |
| super(OCR, self).__init__() |
| self.max_len = max_len |
| self.dictionary = dictionary |
| self.dict_size = len(dictionary) |
| n_decoders = 4 |
| embd_dim = 320 |
| nhead = 4 |
| |
| self.backbone = ConvNext_FeatureExtractor(48, 3, embd_dim) |
| self.encoders = nn.ModuleList() |
| self.decoders = nn.ModuleList() |
| for i in range(4) : |
| encoder = nn.TransformerEncoderLayer(embd_dim, nhead, dropout = 0, batch_first = True, norm_first = True) |
| encoder.self_attn = XposMultiheadAttention(embd_dim, nhead, self_attention = True) |
| encoder.forward = transformer_encoder_forward |
| self.encoders.append(encoder) |
| self.encoders.forward = self.encoder_forward |
|
|
| for i in range(5) : |
| decoder = nn.TransformerDecoderLayer(embd_dim, nhead, dropout = 0, batch_first = True, norm_first = True) |
| decoder.self_attn = XposMultiheadAttention(embd_dim, nhead, self_attention = True) |
| decoder.multihead_attn = XposMultiheadAttention(embd_dim, nhead, encoder_decoder_attention = True) |
| self.decoders.append(decoder) |
| self.decoders.forward = self.decoder_forward |
|
|
| self.embd = nn.Embedding(self.dict_size, embd_dim) |
| self.pred1 = nn.Sequential(nn.Linear(embd_dim, embd_dim), nn.GELU(), nn.Dropout(0.15)) |
| self.pred = nn.Linear(embd_dim, self.dict_size) |
| self.pred.weight = self.embd.weight |
| self.color_pred1 = nn.Sequential(nn.Linear(embd_dim, 64), nn.ReLU()) |
| self.color_pred_fg = nn.Linear(64, 3) |
| self.color_pred_bg = nn.Linear(64, 3) |
| self.color_pred_fg_ind = nn.Linear(64, 2) |
| self.color_pred_bg_ind = nn.Linear(64, 2) |
|
|
| def forward(self, |
| img: torch.FloatTensor, |
| char_idx: torch.LongTensor, |
| decoder_mask: torch.BoolTensor, |
| encoder_mask: torch.BoolTensor |
| ): |
| memory = self.backbone(img) |
| memory = einops.rearrange(memory, 'N C 1 W -> N W C') |
| for layer in self.encoders : |
| memory = layer(memory, src_key_padding_mask = encoder_mask) |
| N, L = char_idx.shape |
| char_embd = self.embd(char_idx) |
| |
| casual_mask = generate_square_subsequent_mask(L).to(img.device) |
| decoded = char_embd |
| for layer in self.decoders : |
| decoded = layer(decoded, memory, tgt_mask = casual_mask, tgt_key_padding_mask = decoder_mask, memory_key_padding_mask = encoder_mask) |
| |
| pred_char_logits = self.pred(self.pred1(decoded)) |
| color_feats = self.color_pred1(decoded) |
| return pred_char_logits, \ |
| self.color_pred_fg(color_feats), \ |
| self.color_pred_bg(color_feats), \ |
| self.color_pred_fg_ind(color_feats), \ |
| self.color_pred_bg_ind(color_feats) |
|
|
| def infer_beam_batch(self, img: torch.FloatTensor, img_widths: List[int], beams_k: int = 5, start_tok = 1, end_tok = 2, pad_tok = 0, max_finished_hypos: int = 2, max_seq_length = 384): |
| N, C, H, W = img.shape |
| assert H == 48 and C == 3 |
| memory = self.backbone(img) |
| memory = einops.rearrange(memory, 'N C 1 W -> N W C') |
| valid_feats_length = [(x + 3) // 4 + 2 for x in img_widths] |
| input_mask = torch.zeros(N, memory.size(1), dtype = torch.bool).to(img.device) |
| for i, l in enumerate(valid_feats_length): |
| input_mask[i, l:] = True |
| for layer in self.encoders : |
| memory = layer(layer, src = memory, src_key_padding_mask = input_mask) |
| hypos = [Hypothesis(img.device, start_tok, end_tok, pad_tok, i, len(self.decoders), 320) for i in range(N)] |
| |
| decoded = next_token_batch(hypos, memory, input_mask, self.decoders, self.embd) |
| |
| pred_char_logprob = self.pred(self.pred1(decoded)).log_softmax(-1) |
| |
| pred_chars_values, pred_chars_index = torch.topk(pred_char_logprob, beams_k, dim = 1) |
| new_hypos: List[Hypothesis] = [] |
| finished_hypos = defaultdict(list) |
| for i in range(N): |
| for k in range(beams_k): |
| new_hypos.append(hypos[i].extend(pred_chars_index[i, k], pred_chars_values[i, k])) |
| hypos = new_hypos |
| for ixx in range(max_seq_length): |
| |
| decoded = next_token_batch(hypos, memory, torch.stack([input_mask[hyp.memory_idx] for hyp in hypos]) , self.decoders, self.embd) |
| |
| pred_char_logprob = self.pred(self.pred1(decoded)).log_softmax(-1) |
| |
| pred_chars_values, pred_chars_index = torch.topk(pred_char_logprob, beams_k, dim = 1) |
| hypos_per_sample = defaultdict(list) |
| h: Hypothesis |
| for i, h in enumerate(hypos): |
| for k in range(beams_k): |
| hypos_per_sample[h.memory_idx].append(h.extend(pred_chars_index[i, k], pred_chars_values[i, k])) |
| hypos = [] |
| |
| for i in hypos_per_sample.keys(): |
| cur_hypos: List[Hypothesis] = hypos_per_sample[i] |
| cur_hypos = sorted(cur_hypos, key = lambda a: a.sort_key())[: beams_k + 1] |
| |
| to_added_hypos = [] |
| sample_done = False |
| for h in cur_hypos: |
| if h.seq_end(): |
| finished_hypos[i].append(h) |
| if len(finished_hypos[i]) >= max_finished_hypos: |
| sample_done = True |
| break |
| else: |
| if len(to_added_hypos) < beams_k: |
| to_added_hypos.append(h) |
| if not sample_done: |
| hypos.extend(to_added_hypos) |
| if len(hypos) == 0: |
| break |
| |
| for i in range(N): |
| if i not in finished_hypos: |
| cur_hypos: List[Hypothesis] = hypos_per_sample[i] |
| cur_hypo = sorted(cur_hypos, key = lambda a: a.sort_key())[0] |
| finished_hypos[i].append(cur_hypo) |
| assert len(finished_hypos) == N |
| result = [] |
| for i in range(N): |
| cur_hypos = finished_hypos[i] |
| cur_hypo = sorted(cur_hypos, key = lambda a: a.sort_key())[0] |
| decoded = cur_hypo.output() |
| color_feats = self.color_pred1(decoded) |
| fg_pred, bg_pred, fg_ind_pred, bg_ind_pred = \ |
| self.color_pred_fg(color_feats), \ |
| self.color_pred_bg(color_feats), \ |
| self.color_pred_fg_ind(color_feats), \ |
| self.color_pred_bg_ind(color_feats) |
| result.append((cur_hypo.out_idx[1:], cur_hypo.prob(), fg_pred[0], bg_pred[0], fg_ind_pred[0], bg_ind_pred[0])) |
| return result |
|
|
| def infer_beam_batch_tensor(self, img: torch.FloatTensor, img_widths: List[int], beams_k: int = 5, start_tok = 1, end_tok = 2, pad_tok = 0, max_finished_hypos: int = 2, max_seq_length = 384): |
| N, C, H, W = img.shape |
| assert H == 48 and C == 3 |
|
|
|
|
| memory = self.backbone(img) |
| memory = einops.rearrange(memory, 'N C 1 W -> N W C') |
| valid_feats_length = [(x + 3) // 4 + 2 for x in img_widths] |
| input_mask = torch.zeros(N, memory.size(1), dtype = torch.bool).to(img.device) |
| |
| for i, l in enumerate(valid_feats_length): |
| input_mask[i, l:] = True |
| memory = self.encoders(memory, input_mask) |
|
|
|
|
| out_idx = torch.full((N, 1), start_tok, dtype=torch.long, device=img.device) |
| cached_activations = torch.zeros(N, len(self.decoders)+1, max_seq_length, 320, device=img.device) |
| log_probs = torch.zeros(N, 1, device=img.device) |
| idx_embedded = self.embd(out_idx[:, -1:]) |
|
|
|
|
| decoded, cached_activations = self.decoders(idx_embedded, cached_activations, memory, input_mask, 0) |
| pred_char_logprob = self.pred(self.pred1(decoded)).log_softmax(-1) |
| pred_chars_values, pred_chars_index = torch.topk(pred_char_logprob, beams_k, dim = 1) |
|
|
|
|
| out_idx = torch.cat([out_idx.unsqueeze(1).expand(-1, beams_k, -1), pred_chars_index.unsqueeze(-1)], dim=-1).reshape(-1, 2) |
| log_probs = pred_chars_values.view(-1, 1) |
| memory = memory.repeat_interleave(beams_k, dim=0) |
| input_mask = input_mask.repeat_interleave(beams_k, dim=0) |
| cached_activations = cached_activations.repeat_interleave(beams_k, dim=0) |
| batch_index = torch.arange(N).repeat_interleave(beams_k, dim=0).to(img.device) |
|
|
|
|
| finished_hypos = defaultdict(list) |
| N_remaining = N |
|
|
|
|
| for step in range(1, max_seq_length): |
| idx_embedded = self.embd(out_idx[:, -1:]) |
| decoded, cached_activations = self.decoders(idx_embedded, cached_activations, memory, input_mask, step) |
| pred_char_logprob = self.pred(self.pred1(decoded)).log_softmax(-1) |
| pred_chars_values, pred_chars_index = torch.topk(pred_char_logprob, beams_k, dim=1) |
|
|
|
|
| finished = out_idx[:, -1] == end_tok |
| pred_chars_values[finished] = 0 |
| pred_chars_index[finished] = end_tok |
|
|
|
|
| |
| new_out_idx = out_idx.unsqueeze(1).expand(-1, beams_k, -1) |
| new_out_idx = torch.cat([new_out_idx, pred_chars_index.unsqueeze(-1)], dim=-1) |
| new_out_idx = new_out_idx.view(-1, step + 2) |
| new_log_probs = log_probs.unsqueeze(1).expand(-1, beams_k, -1) + pred_chars_values.unsqueeze(-1) |
| new_log_probs = new_log_probs.view(-1, 1) |
|
|
|
|
| |
| new_out_idx = new_out_idx.view(N_remaining, -1, step + 2) |
| new_log_probs = new_log_probs.view(N_remaining, -1) |
| batch_topk_log_probs, batch_topk_indices = new_log_probs.topk(beams_k, dim=1) |
| |
| |
| expanded_topk_indices = batch_topk_indices.unsqueeze(-1).expand(-1, -1, new_out_idx.shape[-1]) |
| out_idx = torch.gather(new_out_idx, 1, expanded_topk_indices).reshape(-1, step + 2) |
| log_probs = batch_topk_log_probs.view(-1, 1) |
|
|
|
|
| |
| finished = (out_idx[:, -1] == end_tok) |
| finished = finished.view(N_remaining, beams_k) |
| finished_counts = finished.sum(dim=1) |
| finished_batch_indices = (finished_counts >= max_finished_hypos).nonzero(as_tuple=False).squeeze() |
|
|
|
|
| if finished_batch_indices.numel() == 0: |
| continue |
|
|
|
|
| if finished_batch_indices.dim() == 0: |
| finished_batch_indices = finished_batch_indices.unsqueeze(0) |
| |
| for idx in finished_batch_indices: |
| batch_log_probs = batch_topk_log_probs[idx] |
| best_beam_idx = batch_log_probs.argmax() |
| finished_hypos[batch_index[beams_k * idx].item()] = \ |
| out_idx[idx * beams_k + best_beam_idx], \ |
| torch.exp(batch_log_probs[best_beam_idx]).item(), \ |
| cached_activations[idx * beams_k + best_beam_idx] |
|
|
|
|
| remaining_indexs = [] |
| for i in range(N_remaining): |
| if i not in finished_batch_indices: |
| for j in range(beams_k): |
| remaining_indexs.append(i * beams_k + j) |
|
|
|
|
| if not remaining_indexs: |
| break |
|
|
|
|
| N_remaining = int(len(remaining_indexs) / beams_k) |
| out_idx = out_idx.index_select(0, torch.tensor(remaining_indexs, device=img.device)) |
| log_probs = log_probs.index_select(0, torch.tensor(remaining_indexs, device=img.device)) |
| memory = memory.index_select(0, torch.tensor(remaining_indexs, device=img.device)) |
| cached_activations = cached_activations.index_select(0, torch.tensor(remaining_indexs, device=img.device)) |
| input_mask = input_mask.index_select(0, torch.tensor(remaining_indexs, device=img.device)) |
| batch_index = batch_index.index_select(0, torch.tensor(remaining_indexs, device=img.device)) |
|
|
|
|
| |
| assert len(finished_hypos) == N |
|
|
|
|
| |
| result = [] |
| for i in range(N): |
| final_idx, prob, decoded = finished_hypos[i] |
| color_feats = self.color_pred1(decoded[-1].unsqueeze(0)) |
| fg_pred, bg_pred, fg_ind_pred, bg_ind_pred = \ |
| self.color_pred_fg(color_feats), \ |
| self.color_pred_bg(color_feats), \ |
| self.color_pred_fg_ind(color_feats), \ |
| self.color_pred_bg_ind(color_feats) |
| result.append((final_idx[1:], prob, fg_pred[0], bg_pred[0], fg_ind_pred[0], bg_ind_pred[0])) |
|
|
|
|
| return result |
|
|
| def encoder_forward(self, memory, encoder_mask): |
| for layer in self.encoders : |
| memory = layer(layer, src = memory, src_key_padding_mask = encoder_mask) |
| return memory |
|
|
| def decoder_forward( |
| self, |
| embd: torch.Tensor, |
| cached_activations: torch.Tensor, |
| memory: torch.Tensor, |
| memory_mask: torch.BoolTensor, |
| step: int |
| ): |
|
|
| layer: nn.TransformerDecoderLayer |
| tgt = embd |
|
|
| for l, layer in enumerate(self.decoders): |
| combined_activations = cached_activations[:, l, :step, :] |
| combined_activations = torch.cat([combined_activations, tgt], dim=1) |
| cached_activations[:, l, step, :] = tgt.squeeze(1) |
|
|
| |
| tgt = tgt + layer.self_attn(layer.norm1(tgt), layer.norm1(combined_activations), layer.norm1(combined_activations), q_offset=step)[0] |
| tgt = tgt + layer.multihead_attn(layer.norm2(tgt), memory, memory, key_padding_mask=memory_mask, q_offset=step)[0] |
| tgt = tgt + layer._ff_block(layer.norm3(tgt)) |
|
|
| cached_activations[:, l+1, step, :] = tgt.squeeze(1) |
|
|
| return tgt.squeeze_(1), cached_activations |
|
|
| import numpy as np |
|
|
| def convert_pl_model(filename: str) : |
| sd = torch.load(filename, map_location = 'cpu')['state_dict'] |
| sd2 = {} |
| for k, v in sd.items() : |
| k: str |
| k = k.removeprefix('model.') |
| sd2[k] = v |
| return sd2 |
|
|
| def test_LocalViT_FeatureExtractor() : |
| net = ConvNext_FeatureExtractor(48, 3, 320) |
| inp = torch.randn(2, 3, 48, 512) |
| out = net(inp) |
| print(out.shape) |
|
|
| def test_infer() : |
| with open('alphabet-all-v7.txt', 'r') as fp : |
| dictionary = [s[:-1] for s in fp.readlines()] |
| model = OCR(dictionary, 32) |
| model.eval() |
| sd = convert_pl_model('epoch=0-step=13000.ckpt') |
| model.load_state_dict(sd) |
| model_parameters = filter(lambda p: p.requires_grad, model.parameters()) |
| params = sum([np.prod(p.size()) for p in model_parameters]) |
| print(params) |
|
|
| img = cv2.cvtColor(cv2.imread('test3.png'), cv2.COLOR_BGR2RGB) |
| ratio = img.shape[1] / float(img.shape[0]) |
| new_w = int(round(ratio * 48)) |
| |
| img = cv2.resize(img, (new_w, 48), interpolation=cv2.INTER_AREA) |
|
|
| img_torch = einops.rearrange((torch.from_numpy(img) / 127.5 - 1.0), 'h w c -> 1 c h w') |
|
|
| with torch.no_grad() : |
| idx, prob, fg_pred, bg_pred, fg_ind_pred, bg_ind_pred = model.infer_beam_batch_tensor(img_torch, [new_w], 5, max_seq_length = 32)[0] |
| txt = '' |
| for i in idx : |
| txt += dictionary[i] |
| print(txt, prob) |
| for chid, fg, bg, fg_ind, bg_ind in zip(idx, fg_pred[0], bg_pred[0], fg_ind_pred[0], bg_ind_pred[0]) : |
| has_fg = (fg_ind[1] > fg_ind[0]).item() |
| has_bg = (bg_ind[1] > bg_ind[0]).item() |
| if has_fg : |
| fg = np.clip((fg * 255).numpy(), 0, 255) |
| if has_bg : |
| bg = np.clip((bg * 255).numpy(), 0, 255) |
| print(f'{dictionary[chid]} {fg if has_fg else "None"} {bg if has_bg else "None"}') |
|
|
| if __name__ == "__main__": |
| test_infer() |
|
|