import torch import torch.nn as nn import timm import torch.nn.functional as F mocov3_std = torch.tensor([0.0365, 0.0384, 0.0333, 0.0364, 0.0177, 0.0388, 0.0418, 0.0400, 0.0347, 0.0327, 0.0478, 0.0385, 0.0384, 0.0396, 0.0361, 0.0347, 0.0443, 0.0342, 0.0383, 0.0374, 0.0365, 0.0453, 0.0352, 0.0315, 0.0384, 0.0534, 0.0374, 0.0358, 0.0355, 0.0349, 0.0350, 0.0392, 0.0360, 0.0369, 0.0356, 0.0332, 0.0372, 0.0349, 0.0358, 0.0332, 0.0352, 0.0387, 0.0328, 0.0358, 0.0381, 0.0373, 0.0359, 0.0326, 0.0342, 0.0338, 0.0347, 0.0725, 0.0400, 0.0345, 0.0377, 0.0376, 0.0368, 0.0339, 0.0371, 0.0341, 0.0380, 0.0353, 0.0350, 0.0389, 0.0363, 0.0347, 0.0363, 0.0363, 0.0354, 0.0354, 0.0369, 0.0538, 0.0358, 0.0384, 0.0339, 0.0362, 0.0354, 0.0381, 0.0357, 0.0370, 0.0349, 0.0394, 0.0355, 0.0344, 0.0372, 0.0379, 0.0361, 0.0376, 0.0344, 0.0354, 0.0309, 0.0360, 0.0382, 0.0349, 0.0386, 0.0375, 0.0344, 0.0325, 0.0354, 0.0392, 0.0329, 0.0361, 0.0358, 0.0364, 0.0361, 0.0345, 0.0361, 0.0347, 0.0384, 0.0399, 0.0328, 0.0511, 0.0371, 0.0400, 0.0343, 0.0375, 0.0388, 0.0417, 0.0373, 0.0392, 0.0352, 0.0483, 0.0374, 0.0373, 0.0347, 0.0357, 0.0359, 0.0375, 0.0361, 0.0646, 0.0412, 0.0375, 0.0347, 0.0353, 0.0311, 0.0371, 0.0352, 0.0376, 0.0372, 0.0411, 0.0349, 0.0338, 0.0398, 0.0341, 0.0386, 0.0348, 0.0383, 0.0366, 0.0379, 0.0336, 0.0343, 0.0357, 0.0341, 0.0348, 0.0331, 0.0382, 0.0374, 0.0355, 0.0378, 0.0336, 0.0376, 0.0362, 0.0347, 0.0336, 0.0317, 0.0351, 0.0329, 0.0736, 0.0382, 0.0388, 0.0353, 0.0357, 0.0381, 0.0406, 0.0390, 0.0349, 0.0392, 0.0341, 0.0358, 0.0398, 0.0335, 0.0391, 0.0385, 0.0342, 0.0345, 0.0347, 0.0386, 0.0348, 0.0357, 0.0345, 0.0357, 0.0368, 0.0394, 0.0391, 0.0397, 0.0328, 0.0366, 0.0351, 0.0395, 0.0418, 0.0389, 0.0397, 0.0360, 0.0407, 0.0697, 0.1027, 0.0411, 0.0194, 0.0363, 0.0398, 0.0371, 0.0362, 0.0372, 0.0310, 0.0338, 0.0425, 0.0355, 0.0462, 0.0662, 0.0540, 0.0347, 0.0671, 0.0341, 0.0348, 0.0375, 0.0374, 0.0334, 0.0372, 0.0346, 0.0376, 0.0372, 0.0412, 0.0353, 0.0411, 0.0333, 0.0346, 0.0360, 0.0397, 0.0313, 0.0369, 0.0354, 0.0311, 0.0395, 0.0510, 0.0361, 0.0366, 0.0364, 0.0392, 0.0405, 0.0348, 0.0339, 0.0509, 0.0379, 0.0365, 0.0790, 0.0370, 0.0355, 0.0350, 0.0347, 0.0398, 0.0335, 0.0357, 0.0356, 0.0311, 0.0380, 0.0338, 0.0385, 0.0358, 0.0360, 0.0355, 0.0407, 0.0383, 0.0387, 0.0344, 0.0365, 0.0307, 0.0383, 0.0365, 0.0369, 0.0343, 0.0341, 0.0363, 0.0387, 0.0361, 0.0384, 0.0362, 0.0440, 0.0352, 0.0381, 0.0357, 0.0342, 0.0379, 0.0348, 0.0394, 0.0380, 0.0351, 0.0357, 0.0334, 0.0357, 0.0368, 0.0356, 0.0406, 0.0365, 0.0380, 0.0384, 0.0388, 0.0354, 0.0355, 0.0317, 0.0368, 0.0377, 0.0353, 0.0345, 0.0373, 0.0348, 0.1193, 0.0546, 0.0362, 0.0352, 0.0354, 0.0381, 0.0387, 0.0455, 0.0358, 0.0338, 0.0345, 0.0369, 0.0342, 0.0334, 0.0384, 0.0405, 0.0361, 0.0353, 0.0364, 0.0347, 0.0335, 0.0430, 0.0375, 0.0371, 0.0374, 0.0350, 0.0358, 0.0325, 0.0359, 0.0817, 0.0406, 0.0356, 0.0358, 0.0349, 0.0385, 0.0323, 0.0361, 0.0372, 0.0405, 0.0346, 0.0349, 0.0361, 0.0371, 0.0387, 0.0402, 0.0373, 0.0370, 0.0368, 0.0369, 0.0337, 0.0353, 0.0344, 0.0361, 0.0352, 0.0369, 0.0357, 0.0370, 0.0359, 0.0361, 0.0335, 0.0351, 0.0351, 0.0365, 0.0401, 0.0336, 0.0335, 0.0357, 0.0386, 0.0374, 0.0357, 0.0376, 0.0369, 0.0369, 0.1032, 0.0365, 0.0823, 0.0367, 0.0459, 0.0823, 0.0397, 0.0364, 0.0343, 0.0357, 0.0401, 0.0357, 0.0388, 0.0345, 0.0334, 0.0359, 0.0331, 0.0362, 0.0388, 0.0586, 0.0402, 0.0385, 0.0373, 0.0368, 0.0332, 0.0357, 0.0366, 0.0378, 0.0343, 0.0348, 0.0338, 0.0381, 0.0359, 0.0342, 0.0385, 0.0347, 0.0359, 0.0396, 0.0366, 0.0374, 0.0305, 0.0689, 0.0389, 0.0359, 0.0373, 0.0359, 0.0377, 0.0367, 0.0367, 0.0454, 0.0364, 0.0327, 0.0347, 0.0406, 0.0374, 0.0360, 0.0336, 0.0363, 0.0364, 0.0413, 0.0364, 0.0348, 0.0591, 0.0390, 0.0327, 0.0424, 0.0369, 0.0356, 0.0350, 0.0369, 0.0362, 0.0346, 0.0479, 0.0319, 0.0370, 0.0406, 0.0371, 0.0387, 0.0346, 0.0389, 0.0394, 0.0363, 0.0331, 0.0371, 0.0391, 0.0365, 0.0389, 0.0362, 0.0380, 0.0378, 0.0369, 0.0353, 0.0365, 0.0363, 0.0374, 0.0362, 0.0642, 0.0399, 0.0349, 0.0349, 0.0381, 0.0312, 0.0544, 0.0330, 0.0346, 0.0359, 0.0428, 0.0365, 0.0380, 0.0370, 0.0377, 0.0378, 0.0377, 0.0370, 0.0364, 0.0365, 0.0368, 0.0375, 0.0416, 0.0382, 0.0348, 0.0373, 0.0353, 0.0372, 0.0357, 0.0338, 0.0395, 0.0358, 0.0359, 0.0382, 0.0352, 0.0356, 0.0340, 0.0396, 0.0424, 0.0387, 0.0304, 0.0364, 0.0401, 0.0511, 0.0392, 0.0370, 0.0363, 0.0405, 0.0341, 0.0348, 0.0334, 0.0346, 0.0349, 0.0392, 0.0370, 0.0406, 0.0372, 0.0369, 0.0364, 0.0357, 0.0385, 0.0387, 0.0366, 0.0307, 0.0350, 0.0357, 0.0422, 0.0404, 0.0334, 0.0345, 0.0532, 0.0368, 0.0333, 0.0374, 0.0371, 0.0381, 0.0397, 0.0384, 0.0354, 0.0353, 0.0343, 0.0415, 0.0368, 0.0351, 0.0398, 0.0395, 0.0365, 0.0370, 0.0367, 0.0370, 0.0362, 0.0349, 0.0372, 0.0327, 0.0367, 0.0373, 0.0426, 0.0348, 0.0349, 0.0473, 0.0331, 0.0365, 0.0376, 0.0352, 0.0366, 0.0410, 0.0381, 0.0373, 0.0407, 0.0331, 0.0364, 0.0497, 0.0375, 0.0378, 0.0349, 0.0413, 0.0348, 0.0379, 0.0354, 0.0364, 0.0347, 0.0356, 0.0347, 0.0383, 0.0714, 0.0392, 0.0452, 0.0353, 0.0373, 0.0361, 0.0358, 0.0348, 0.0362, 0.0377, 0.0350, 0.0354, 0.0365, 0.0360, 0.0367, 0.0366, 0.0358, 0.0357, 0.0357, 0.0508, 0.0368, 0.0353, 0.0419, 0.0344, 0.0380, 0.0338, 0.0363, 0.0370, 0.0355, 0.0358, 0.0367, 0.0375, 0.0375, 0.0559, 0.0361, 0.0378, 0.0381, 0.0343, 0.0379, 0.0390, 0.0396, 0.0360, 0.0388, 0.0351, 0.0362, 0.0351, 0.0357, 0.0349, 0.0336, 0.0371, 0.0344, 0.0358, 0.0354, 0.0382, 0.0386, 0.0406, 0.0834, 0.0361, 0.0360, 0.0361, 0.0351, 0.0379, 0.0355, 0.0390, 0.0364, 0.0351, 0.0374, 0.0436, 0.0375, 0.0363, 0.0353, 0.0388, 0.0355, 0.0348, 0.0364, 0.0325, 0.0340, 0.0343, 0.0389, 0.0358, 0.0348, 0.0349, 0.0373, 0.0361, 0.0364, 0.0367, 0.0373, 0.0377, 0.0322, 0.0379, 0.0333, 0.0442, 0.0389, 0.0324, 0.0367, 0.0356, 0.0345, 0.0393, 0.0349, 0.0450, 0.0382, 0.0376, 0.0463, 0.0363, 0.0328, 0.0356, 0.0379, 0.0360, 0.0342, 0.0371, 0.0356, 0.0373, 0.0355, 0.0367, 0.0313, 0.0425, 0.0366, 0.0352, 0.0366, 0.0363, 0.0323, 0.0328, 0.0335, 0.0337, 0.0402, 0.0369, 0.0390, 0.0363, 0.0416, 0.0592, 0.0343, 0.0338, 0.0371, 0.0722, 0.0449, 0.0350, 0.0356, 0.0352, 0.0361, 0.0366, 0.0362, 0.0463, 0.0347, 0.0400, 0.0327, 0.0362, 0.0375, 0.0466, 0.0341, 0.0332, 0.0325, 0.0369, 0.0326, 0.0373, 0.0374, 0.0367, 0.0365, 0.0344, 0.0398, 0.0378]) mocov3_mean = torch.tensor([-4.9909e-03, 5.8531e-02, -8.0204e-02, 1.4484e-02, 6.5256e-04, 3.1926e-02, 5.2389e-02, -4.6138e-02, -2.9104e-02, -1.0310e-03, 1.4314e-02, 4.3464e-02, 5.4860e-02, -3.8034e-03, 9.6628e-02, 6.7566e-02, -2.0503e-01, -5.7046e-02, -8.4732e-02, -5.1926e-02, 2.8064e-02, -7.4545e-02, -3.0411e-02, -2.1032e-02, 1.0223e-02, -3.9128e-02, -1.0685e-01, -4.2874e-02, 7.4012e-02, -8.5295e-02, -5.1053e-02, 1.1215e-01, -3.4985e-02, -1.9459e-02, -5.4159e-02, -3.3352e-02, -2.7664e-02, 6.8211e-02, -5.2040e-02, 1.4412e-02, -5.8436e-02, 2.2623e-02, 1.6369e-02, -2.6669e-02, 7.5853e-03, -1.7022e-02, 1.9521e-02, 1.7904e-02, -1.7904e-02, -5.8781e-02, -5.1144e-02, -5.0436e-03, -2.6308e-02, 3.3595e-03, 2.5913e-02, 2.7867e-03, -9.1458e-02, -4.5019e-02, -3.1314e-02, -4.4559e-02, -5.5143e-02, 1.8014e-02, 2.7575e-02, -4.7217e-02, 4.3467e-02, -8.2260e-02, -6.7334e-03, -5.6354e-02, 7.9308e-02, -2.9664e-02, -1.8751e-02, -9.8325e-02, 7.9536e-02, -1.3846e-02, 3.8479e-02, -2.6752e-02, 7.3832e-02, -3.5585e-03, 9.6148e-02, 2.4930e-02, -3.7335e-04, -3.6863e-02, -2.6756e-02, 4.9271e-02, -2.8841e-02, -3.0766e-03, 8.6419e-02, 6.1747e-02, -6.5190e-02, 5.2677e-02, -1.5961e-02, -8.7862e-03, 5.4241e-02, 5.0302e-02, -7.1608e-02, 3.5493e-02, -2.4699e-02, -5.9670e-02, 8.6370e-03, -3.0705e-02, -8.9275e-03, -1.3236e-02, -1.0522e-01, -5.8801e-02, 4.5025e-03, -1.9658e-02, -3.5878e-02, 7.4032e-03, -1.2370e-02, -5.7225e-02, 6.1609e-02, 4.5315e-03, -4.6809e-02, 6.3947e-02, 1.8732e-02, -2.0228e-02, -6.8604e-03, 2.5951e-03, -9.4610e-03, 3.7075e-02, -3.4487e-02, 8.2085e-02, 1.9499e-03, 1.8524e-02, -2.0771e-02, 4.1893e-02, -5.9360e-02, 2.2461e-02, -1.3253e-02, -2.4474e-02, -1.0797e-01, -3.7946e-02, -2.8403e-02, 4.2671e-02, -6.6034e-02, 4.8225e-05, 5.0553e-02, -1.6483e-02, -9.2907e-02, -5.0918e-02, -4.4737e-02, 1.0900e-03, -2.2239e-02, -1.5004e-01, 1.5945e-02, 1.0725e-01, 5.8067e-02, 4.5711e-03, 2.8622e-02, 6.3638e-04, -9.8273e-03, 4.3623e-02, 8.0471e-02, 6.2474e-02, 5.5382e-02, -3.9220e-02, -3.7855e-02, 1.7026e-02, 6.4659e-02, -4.5883e-03, 8.0370e-02, 1.5961e-02, 1.9753e-02, -9.2319e-03, 4.0418e-02, 4.7325e-03, 4.7597e-02, 1.9679e-03, 8.5419e-02, 3.2275e-02, 5.3665e-02, -3.7527e-02, -1.0371e-01, -1.0602e-02, -6.3089e-02, 2.1686e-02, -2.1665e-03, -4.0568e-02, -5.4864e-02, 1.2588e-02, -4.0154e-02, -1.0553e-01, -1.7084e-02, 5.4653e-02, 1.3699e-01, -2.0908e-02, 3.0141e-02, -2.8898e-02, -6.7819e-04, -1.5839e-02, 1.6851e-02, 3.5577e-02, -4.6329e-03, -6.0051e-02, 2.6143e-02, 1.7365e-02, -1.9256e-02, 3.1024e-02, 5.1721e-02, -8.4824e-02, 6.7965e-02, -9.0949e-02, 4.3225e-03, -7.1089e-02, 9.0626e-02, -1.1133e-01, -1.8509e-02, 5.5376e-02, 3.3383e-02, -2.9887e-02, -5.2169e-02, 2.7724e-02, 1.2762e-02, 6.4547e-02, 6.4587e-02, 1.9503e-02, 1.2702e-02, 1.5315e-02, 1.6836e-02, -1.3361e-01, -9.1265e-02, -3.5769e-02, 4.2201e-02, 1.1958e-02, 5.4143e-03, -4.0032e-02, 8.6459e-02, 6.5958e-02, 1.5095e-02, -8.1063e-02, 2.3027e-02, 2.4872e-02, 2.2488e-02, 9.4502e-02, -1.9964e-02, 1.3892e-02, 6.8407e-02, -1.8256e-02, 2.1523e-02, -2.7407e-02, -1.0688e-02, -1.6652e-02, -1.9957e-02, 5.7456e-02, -1.1604e-01, -5.1513e-02, -1.1086e-01, 7.0534e-02, -1.4484e-02, 4.3226e-03, 1.0349e-01, -3.6197e-03, 8.9538e-02, -1.5885e-02, -3.8066e-02, 5.3550e-02, 6.5762e-02, -9.3377e-02, 1.1694e-03, -2.6036e-02, -4.9501e-02, 7.8390e-02, -2.3242e-03, -5.7216e-02, 7.1508e-03, -4.0947e-02, 1.2399e-01, -3.8855e-03, -1.7977e-02, -1.2340e-01, -5.2535e-02, 3.4631e-02, 1.9590e-02, 6.8397e-03, -2.1967e-02, 2.6737e-02, -2.3443e-02, -8.6921e-03, -4.3219e-02, -4.7690e-02, 1.5660e-02, 6.5133e-02, 5.2079e-02, -6.9453e-02, -5.0886e-02, -3.9676e-02, -1.0395e-02, -5.7200e-02, 4.2374e-02, 8.4635e-02, 5.6200e-02, 1.9147e-03, -2.8570e-02, 5.8081e-02, -4.9870e-02, 8.7587e-02, -2.1561e-02, -3.9530e-02, 3.1023e-02, -3.2769e-02, -3.1827e-02, 5.4912e-02, -6.4331e-02, 2.7848e-02, -1.4435e-01, -1.3247e-02, 4.1248e-02, 2.4548e-03, -6.6767e-02, -2.2580e-02, -4.4151e-02, -5.4278e-02, 1.0243e-02, 4.1607e-02, -3.6166e-02, 8.6948e-02, 6.7075e-02, -1.1791e-01, 8.5181e-02, -1.2960e-02, -1.8911e-02, 5.7109e-02, -3.8419e-02, -9.0466e-02, 4.8168e-03, -2.6800e-02, 4.8001e-02, 9.3661e-02, 1.2480e-01, 2.9446e-02, -7.8841e-02, 1.9951e-02, -5.6541e-02, 6.8220e-03, -1.4068e-02, -2.4581e-02, -7.4762e-02, 4.4282e-02, 9.4121e-02, -8.8464e-03, 1.2429e-02, 2.3094e-02, 2.2941e-02, -7.1354e-02, 2.9385e-01, -1.6506e-02, -6.4741e-03, -3.3756e-02, -1.9813e-02, 1.0995e-02, -1.5479e-02, 8.3897e-03, -1.7868e-02, -5.3500e-02, -1.5280e-01, -2.4950e-02, -4.3315e-02, -5.0033e-02, 5.7700e-02, -2.2942e-02, 2.8033e-02, -7.6005e-02, -1.5828e-02, 4.6690e-02, -1.2853e-01, 8.6603e-02, 9.7747e-03, -1.5809e-02, -1.0410e-01, 4.8349e-02, -7.7004e-03, -7.7936e-02, -5.0415e-02, -4.1027e-02, -5.7368e-02, 6.3061e-02, -1.4738e-01, 9.0193e-03, 1.4264e-02, 8.1013e-02, 1.4018e-02, 4.0492e-02, 5.7056e-03, 1.3297e-02, 2.2139e-02, -5.1328e-02, -5.6951e-02, 1.8988e-02, 3.5907e-02, -7.2411e-02, 1.8173e-01, 9.4154e-03, 2.4558e-04, 9.9371e-02, 3.6060e-03, 5.8868e-02, -1.6166e-02, -3.3660e-03, -4.0091e-02, -7.3601e-02, -4.7921e-02, 5.5936e-02, 3.3641e-02, -5.8169e-03, 1.7200e-03, 1.4891e-03, 4.6669e-02, 2.6898e-02, -2.0748e-02, -3.7035e-02, 7.1686e-02, -1.0291e-01, 5.2060e-02, 5.8818e-02, 5.6555e-04, -4.9335e-02, 6.5931e-02, 1.1788e-01, 7.9230e-03, 7.9755e-02, 5.8611e-02, -4.8066e-02, 8.9253e-03, -1.4002e-02, -6.4210e-02, 4.7972e-03, 5.6867e-02, 1.2133e-01, -8.8800e-02, 1.0355e-01, -1.8507e-02, 2.1943e-04, -1.9579e-02, 6.1412e-03, 2.9627e-02, 6.6325e-02, -6.3568e-02, 6.6112e-02, 7.5182e-02, 3.0145e-03, 7.8490e-03, 3.8883e-02, 9.4810e-02, 6.8651e-02, 9.3259e-02, 7.6506e-02, -1.1391e-01, 4.2561e-02, 9.0650e-02, 5.0542e-02, 4.6873e-02, 2.9639e-02, -7.0080e-02, -8.2195e-03, 1.7181e-02, -3.2508e-02, -5.6378e-03, -1.0345e-01, 1.0686e-02, -7.7440e-02, 1.0137e-01, 1.1383e-01, 4.9601e-02, 1.0087e-01, 6.7788e-03, -5.2826e-02, 2.5621e-02, -9.7276e-02, -9.5657e-02, 3.9823e-02, 8.1216e-03, 5.6924e-02, 9.2770e-03, 3.0634e-02, 5.1213e-02, -1.9143e-02, -1.4004e-02, 1.0268e-01, 2.6726e-02, 1.3936e-03, 1.0327e-02, 3.1209e-03, -1.9589e-02, -3.8289e-02, 7.4170e-02, 5.5759e-02, 1.6756e-02, 1.5523e-01, 8.1868e-02, 7.1569e-03, -1.0032e-01, -3.1106e-02, -4.5538e-02, 5.4754e-02, -4.1384e-03, 6.6185e-04, 8.9977e-02, 1.1297e-01, -3.2720e-02, -1.1563e-02, 2.9044e-02, 1.1240e-01, -6.2860e-02, -1.8176e-02, 3.7151e-02, 9.9272e-02, -1.5495e-02, -7.4552e-02, 7.0645e-02, 7.8578e-03, -9.8891e-02, -6.4297e-02, 9.9785e-02, 8.1721e-03, -6.9688e-02, 1.1140e-03, -4.7135e-02, 5.7858e-02, 1.8430e-02, -1.4128e-03, -3.9728e-02, 7.4199e-03, -8.4799e-02, 1.2801e-02, 1.5855e-01, -1.4473e-02, -3.2350e-02, 8.8111e-02, 7.3757e-02, -5.9830e-02, -6.8308e-02, -4.4946e-02, -1.1461e-02, -3.8021e-02, 1.9265e-03, -4.3868e-02, 2.3531e-02, 4.0215e-02, -3.4769e-03, 1.7513e-02, -3.2142e-02, 1.4742e-02, -3.9892e-02, -1.5716e-02, -7.3177e-03, -2.1507e-02, 2.4539e-02, -4.8903e-03, -4.0340e-02, -3.9031e-02, 1.6703e-01, -3.3599e-02, -8.6483e-02, 3.9196e-02, 5.4656e-02, -6.0245e-02, 8.4745e-02, -1.0317e-01, -3.2529e-02, -3.1298e-02, 4.5586e-02, 1.7218e-02, -5.1573e-02, -9.2497e-02, -3.7508e-02, 9.3142e-02, 3.7266e-02, -7.8790e-02, -2.2965e-03, -8.8131e-03, 1.1247e-01, 1.4398e-02, 3.6883e-02, 1.3535e-02, -2.2482e-02, 7.7606e-03, -1.1564e-02, -9.2380e-02, 9.2117e-02, -8.6496e-02, -4.7756e-02, 9.0737e-02, 4.9368e-02, -1.3184e-02, -3.3454e-02, 6.3449e-03, 3.7228e-02, -2.4421e-03, 1.7024e-03, 2.9668e-02, -5.9003e-02, -4.5759e-02, 5.7492e-02, -6.7707e-03, 3.1389e-02, 2.8321e-02, -6.0809e-02, -2.4451e-03, 6.3512e-03, -1.8711e-02, 1.2424e-02, 1.3970e-02, 5.6096e-03, -9.3184e-02, 1.4589e-02, -1.9262e-03, -3.8669e-02, -5.9775e-02, -2.1852e-05, -1.3025e-02, -4.6614e-02, 2.0437e-02, -2.0632e-03, 9.9417e-02, -8.5307e-02, -6.2060e-02, -4.7353e-02, -8.4368e-02, 5.9125e-02, 1.3815e-02, 6.7903e-02, -1.9176e-02, 8.1607e-02, 9.5300e-03, -7.5729e-02, 4.8691e-02, 7.7487e-02, -1.1334e-02, -4.1340e-02, 1.1121e-02, -2.4506e-02, 1.4673e-01, -4.1853e-02, 6.9515e-02, -1.2374e-01, -5.0617e-02, 2.6989e-02, 1.0692e-02, -5.9575e-02, 3.0577e-02, 5.9018e-02, -2.8683e-03, -5.3847e-02, -1.3862e-02, -7.3678e-03, -2.5915e-02, 5.4326e-02, -1.0355e-02, -1.4217e-02, -3.8828e-02, -4.3901e-02, -3.6999e-02, -8.7978e-02, -3.7600e-02, 7.3855e-02, -4.6035e-02, -1.0064e-02, 8.4649e-02, 5.4544e-02, 6.6927e-02, -6.7527e-02, 4.1447e-02, -5.2767e-02, 5.7202e-02, 2.0367e-02, -1.4843e-02, -1.3606e-02, -6.5079e-02, 4.3446e-02, -3.8950e-02, 3.5919e-02, 2.0620e-02, -6.4451e-02, 1.5398e-02, -2.2867e-03, 1.1872e-02, 1.0481e-01, 5.0847e-02, -7.6596e-02, -9.0736e-02, -3.9243e-02, -3.5298e-02, 2.8835e-02, -1.2076e-02, -1.7122e-02, -1.9240e-02, 4.2200e-02, -2.8937e-03, 1.0121e-01, -1.1783e-02, 1.6700e-02, -3.5787e-02, 8.3502e-02, -7.7383e-03, -1.8015e-02, -2.6297e-02, 6.3130e-02, 2.6768e-02, 1.5342e-02, -5.8850e-02, -1.4420e-01, -5.5196e-03, 7.0583e-02, -3.3988e-02, -5.3023e-02, 3.2967e-02, 1.8001e-02, 6.6126e-02, 5.1371e-02, -8.5679e-02, -5.4478e-02, -3.4152e-02, -2.4631e-03, -4.2347e-02, 2.5897e-02, 6.2213e-02, 4.1291e-02, -5.2986e-02, -2.5307e-02, 2.5769e-03, 4.4660e-03, -3.7381e-04, 2.0208e-03, 5.6412e-02, -1.5519e-02, 5.4275e-02, 7.7032e-03, -2.1544e-02, 9.5500e-02, 1.2722e-02, -3.2053e-03, -4.3996e-02, -4.5722e-02, 2.6009e-02, -6.7119e-02, -2.0980e-02, 2.5636e-02, -6.9467e-02, -1.1082e-01, 1.2189e-02, 2.5662e-02, 3.2322e-02, -8.3210e-02, 1.8254e-02, -2.7739e-02, 3.8996e-02, 1.3017e-02, 4.5556e-02, -9.2341e-02, -8.9775e-02, 8.1507e-02, 2.7921e-03, -1.1678e-01, 4.3381e-02, 5.3091e-02, -1.0354e-01, -3.7595e-02, 8.8258e-02, -7.4204e-02, 9.5140e-02, -2.0128e-02, -4.3937e-02, 2.2460e-02, 6.1824e-03, -3.3601e-02, -5.7980e-03, -4.4147e-02, -1.0026e-01]) #https://github.com/huggingface/pytorch-image-models/blob/main/timm/models/vision_transformer.py #https://github.com/huggingface/pytorch-image-models/blob/main/timm/layers/pos_embed.py#L19 # # weight_url = "https://dl.fbaipublicfiles.com/moco-v3/vit-b-300ep/vit-b-300ep.pth.tar" # checkpoint_path = "vit-b-300ep.pth.tar" # if not os.path.exists(checkpoint_path): # print(f"Downloading weights from {weight_url}...") # torch.hub.download_url_to_file(weight_url, checkpoint_path) from typing import Optional, Union, Tuple, List @torch.fx.wrap def resample_abs_pos_embed( posemb: torch.Tensor, new_size: List[int], old_size: Optional[List[int]] = None, num_prefix_tokens: int = 1, interpolation: str = 'bicubic', antialias: bool = True, verbose: bool = False, ): # sort out sizes, assume square if old size not provided num_pos_tokens = posemb.shape[1] num_new_tokens = new_size[0] * new_size[1] + num_prefix_tokens if num_new_tokens == num_pos_tokens and new_size[0] == new_size[1]: return posemb if old_size is None: hw = int(math.sqrt(num_pos_tokens - num_prefix_tokens)) old_size = hw, hw if num_prefix_tokens: posemb_prefix, posemb = posemb[:, :num_prefix_tokens], posemb[:, num_prefix_tokens:] else: posemb_prefix, posemb = None, posemb # do the interpolation embed_dim = posemb.shape[-1] orig_dtype = posemb.dtype posemb = posemb.float() # interpolate needs float32 posemb = posemb.reshape(1, old_size[0], old_size[1], -1).permute(0, 3, 1, 2) posemb = F.interpolate(posemb, size=new_size, mode=interpolation, antialias=antialias) posemb = posemb.permute(0, 2, 3, 1).reshape(1, -1, embed_dim) posemb = posemb.to(orig_dtype) # add back extra (class, etc) prefix tokens if posemb_prefix is not None: posemb = torch.cat([posemb_prefix, posemb], dim=1) return posemb def to_2tuple(x): return x,x def _init_img_size(self, img_size: Union[int, Tuple[int, int]]): assert self.patch_size if img_size is None: return None, None, None img_size = to_2tuple(img_size) grid_size = tuple([s // p for s, p in zip(img_size, self.patch_size)]) num_patches = grid_size[0] * grid_size[1] return img_size, grid_size, num_patches def set_input_size_patchembed( self, img_size: Optional[Union[int, Tuple[int, int]]] = None, patch_size: Optional[Union[int, Tuple[int, int]]] = None, ): new_patch_size = None if patch_size is not None: new_patch_size = to_2tuple(patch_size) if new_patch_size is not None and new_patch_size != self.patch_size: with torch.no_grad(): new_proj = nn.Conv2d( self.proj.in_channels, self.proj.out_channels, kernel_size=new_patch_size, stride=new_patch_size, bias=self.proj.bias is not None, device=self.proj.weight.device, dtype=self.proj.weight.dtype, ) new_proj.weight.copy_(resample_patch_embed(self.proj.weight, new_patch_size, verbose=True)) if self.proj.bias is not None: new_proj.bias.copy_(self.proj.bias) self.proj = new_proj self.patch_size = new_patch_size img_size = img_size or self.img_size if img_size != self.img_size or new_patch_size is not None: self.img_size, self.grid_size, self.num_patches = _init_img_size(self, img_size) def set_input_size( self, img_size: Optional[Union[int, Tuple[int, int]]] = None, patch_size: Optional[Union[int, Tuple[int, int]]] = None, ) -> None: """Update the input image resolution and patch size. Args: img_size: New input resolution, if None current resolution is used. patch_size: New patch size, if None existing patch size is used. """ prev_grid_size = self.patch_embed.grid_size set_input_size_patchembed(self.patch_embed, img_size=img_size, patch_size=patch_size) if self.pos_embed is not None: num_prefix_tokens = 0 if self.no_embed_class else self.num_prefix_tokens num_new_tokens = self.patch_embed.num_patches + num_prefix_tokens if num_new_tokens != self.pos_embed.shape[1]: self.pos_embed = nn.Parameter(resample_abs_pos_embed( self.pos_embed, new_size=self.patch_embed.grid_size, old_size=prev_grid_size, num_prefix_tokens=num_prefix_tokens, verbose=True, )) class MocoV3(nn.Module): def __init__( self, model_ckpt_path: str = '/path/to/latentforcing/mocov3b.pth.tar', match_pixel_norm: float = 0.485, ): super().__init__() self.register_buffer("latent_std", mocov3_std.clone().float()) self.register_buffer("latent_mean", mocov3_mean.clone().float()) self.register_buffer("pixel_std", torch.tensor((0.229, 0.224, 0.225))) self.register_buffer("pixel_mean", torch.tensor((0.485, 0.456, 0.406))) self.match_pixel_norm = match_pixel_norm checkpoint = torch.load(model_ckpt_path, map_location="cpu") state_dict = checkpoint['state_dict'] new_state_dict = {} for k, v in state_dict.items(): if k.startswith("module.base_encoder."): new_k = k.replace("module.base_encoder.", "") new_state_dict[new_k] = v self.mocov3 = timm.create_model("vit_base_patch16_224", num_classes=0) self.mocov3.load_state_dict(new_state_dict, strict=False) set_input_size(self.mocov3, 256) self.mocov3.eval() self.mocov3.requires_grad_(False) @torch.compile() @torch.no_grad() def encode(self, x: torch.Tensor) -> torch.Tensor: # normalize input # x : b c h w x = (x - self.pixel_mean.view(1,3,1,1)) / self.pixel_std.view(1,3,1,1) z = self.mocov3.forward_features(x) # b 1+n d z = z[:,1:] # remove cls z = (z - self.latent_mean.view(1,1,-1)) / self.latent_std.view(1,1,-1) z = z * self.match_pixel_norm z = z.view(-1,16,16,768).permute(0,3,1,2) # b hw d --> b d h w return z