Upload folder using huggingface_hub
Browse files
2026-04-12-154000-flow-warp-ar-curriculum/config.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"in_channels": 24, "channels": [32, 64, 128, 256], "context_len": 8, "model_class": "FlowWarpAttnUNet"}
|
2026-04-12-154000-flow-warp-ar-curriculum/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ac2e2d6dd460bcf391eb6a0d7e17bde8bbbf0b1d82eb6f7aa7402c70855c7f9f
|
| 3 |
+
size 15231294
|
2026-04-12-154000-flow-warp-ar-curriculum/predict.py
ADDED
|
@@ -0,0 +1,59 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""Inference for AR curriculum model + TTA."""
|
| 2 |
+
import json
|
| 3 |
+
import numpy as np
|
| 4 |
+
import torch
|
| 5 |
+
import sys
|
| 6 |
+
sys.path.insert(0, "/home/coder/code")
|
| 7 |
+
from flow_warp_attn_model import FlowWarpAttnUNet
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
def load_model(model_dir: str):
|
| 11 |
+
with open(f"{model_dir}/config.json") as f:
|
| 12 |
+
config = json.load(f)
|
| 13 |
+
model = FlowWarpAttnUNet(in_channels=config["in_channels"], channels=config["channels"])
|
| 14 |
+
sd = torch.load(f"{model_dir}/model.pt", map_location="cpu", weights_only=True)
|
| 15 |
+
sd = {k: v.float() for k, v in sd.items()}
|
| 16 |
+
model.load_state_dict(sd)
|
| 17 |
+
model.eval()
|
| 18 |
+
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 19 |
+
model = model.to(device)
|
| 20 |
+
return {"model": model, "device": device, "context_len": config["context_len"]}
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
def _prepare_input(context_frames, context_len):
|
| 24 |
+
N = len(context_frames)
|
| 25 |
+
if N >= context_len:
|
| 26 |
+
frames = context_frames[-context_len:]
|
| 27 |
+
else:
|
| 28 |
+
pad = np.repeat(context_frames[:1], context_len - N, axis=0)
|
| 29 |
+
frames = np.concatenate([pad, context_frames], axis=0)
|
| 30 |
+
frames_f = frames.astype(np.float32) / 255.0
|
| 31 |
+
frames_f = np.transpose(frames_f, (0, 3, 1, 2))
|
| 32 |
+
context = frames_f.reshape(1, -1, 64, 64)
|
| 33 |
+
last_frame = frames_f[-1:]
|
| 34 |
+
return context, last_frame
|
| 35 |
+
|
| 36 |
+
|
| 37 |
+
def predict_next_frame(model_dict, context_frames: np.ndarray) -> np.ndarray:
|
| 38 |
+
model = model_dict["model"]
|
| 39 |
+
device = model_dict["device"]
|
| 40 |
+
context_len = model_dict["context_len"]
|
| 41 |
+
|
| 42 |
+
ctx, last = _prepare_input(context_frames, context_len)
|
| 43 |
+
with torch.no_grad():
|
| 44 |
+
ctx_t = torch.from_numpy(ctx).to(device)
|
| 45 |
+
last_t = torch.from_numpy(last).to(device)
|
| 46 |
+
pred1, _ = model(ctx_t, last_t)
|
| 47 |
+
|
| 48 |
+
flipped_frames = context_frames[:, :, ::-1, :].copy()
|
| 49 |
+
ctx_f, last_f = _prepare_input(flipped_frames, context_len)
|
| 50 |
+
with torch.no_grad():
|
| 51 |
+
ctx_ft = torch.from_numpy(ctx_f).to(device)
|
| 52 |
+
last_ft = torch.from_numpy(last_f).to(device)
|
| 53 |
+
pred2, _ = model(ctx_ft, last_ft)
|
| 54 |
+
pred2 = pred2.flip(-1)
|
| 55 |
+
|
| 56 |
+
pred = (pred1 + pred2) / 2.0
|
| 57 |
+
pred_np = pred[0].cpu().numpy()
|
| 58 |
+
pred_np = np.transpose(pred_np, (1, 2, 0))
|
| 59 |
+
return (pred_np * 255.0).clip(0, 255).astype(np.uint8)
|
2026-04-12-154000-flow-warp-ar-curriculum/train.log
ADDED
|
@@ -0,0 +1,254 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[15:40:44] Model: FlowWarpAttnUNet, 7,596,742 params
|
| 2 |
+
[15:40:44] === Phase 1: Single-step (80 epochs) ===
|
| 3 |
+
[15:40:49] Train: 44392, Val: 5456
|
| 4 |
+
[15:41:38] P1 Ep 1/80 | Train: 0.112908 | Val: 0.092624 | LR: 3.00e-04
|
| 5 |
+
[15:41:38] -> Saved P1 (val=0.092624)
|
| 6 |
+
[15:42:16] P1 Ep 2/80 | Train: 0.086134 | Val: 0.081807 | LR: 3.00e-04
|
| 7 |
+
[15:42:16] -> Saved P1 (val=0.081807)
|
| 8 |
+
[15:42:41] P1 Ep 3/80 | Train: 0.080104 | Val: 0.077833 | LR: 2.99e-04
|
| 9 |
+
[15:42:41] -> Saved P1 (val=0.077833)
|
| 10 |
+
[15:43:27] P1 Ep 4/80 | Train: 0.076518 | Val: 0.075106 | LR: 2.98e-04
|
| 11 |
+
[15:43:27] -> Saved P1 (val=0.075106)
|
| 12 |
+
[15:44:25] P1 Ep 5/80 | Train: 0.073637 | Val: 0.073171 | LR: 2.97e-04
|
| 13 |
+
[15:44:25] -> Saved P1 (val=0.073171)
|
| 14 |
+
[15:45:17] P1 Ep 6/80 | Train: 0.071194 | Val: 0.071489 | LR: 2.96e-04
|
| 15 |
+
[15:45:17] -> Saved P1 (val=0.071489)
|
| 16 |
+
[15:46:12] P1 Ep 7/80 | Train: 0.069269 | Val: 0.069629 | LR: 2.94e-04
|
| 17 |
+
[15:46:12] -> Saved P1 (val=0.069629)
|
| 18 |
+
[15:47:07] P1 Ep 8/80 | Train: 0.067358 | Val: 0.068144 | LR: 2.93e-04
|
| 19 |
+
[15:47:07] -> Saved P1 (val=0.068144)
|
| 20 |
+
[15:48:01] P1 Ep 9/80 | Train: 0.065801 | Val: 0.066217 | LR: 2.91e-04
|
| 21 |
+
[15:48:01] -> Saved P1 (val=0.066217)
|
| 22 |
+
[15:48:48] P1 Ep 10/80 | Train: 0.064228 | Val: 0.064792 | LR: 2.89e-04
|
| 23 |
+
[15:48:48] -> Saved P1 (val=0.064792)
|
| 24 |
+
[15:49:11] P1 Ep 11/80 | Train: 0.062906 | Val: 0.064103 | LR: 2.86e-04
|
| 25 |
+
[15:49:11] -> Saved P1 (val=0.064103)
|
| 26 |
+
[15:49:47] P1 Ep 12/80 | Train: 0.061461 | Val: 0.063365 | LR: 2.84e-04
|
| 27 |
+
[15:49:47] -> Saved P1 (val=0.063365)
|
| 28 |
+
[15:50:31] P1 Ep 13/80 | Train: 0.060517 | Val: 0.062029 | LR: 2.81e-04
|
| 29 |
+
[15:50:31] -> Saved P1 (val=0.062029)
|
| 30 |
+
[15:51:25] P1 Ep 14/80 | Train: 0.059288 | Val: 0.062736 | LR: 2.78e-04
|
| 31 |
+
[15:52:11] P1 Ep 15/80 | Train: 0.058430 | Val: 0.060402 | LR: 2.75e-04
|
| 32 |
+
[15:52:11] -> Saved P1 (val=0.060402)
|
| 33 |
+
[15:52:35] P1 Ep 16/80 | Train: 0.057286 | Val: 0.058811 | LR: 2.71e-04
|
| 34 |
+
[15:52:35] -> Saved P1 (val=0.058811)
|
| 35 |
+
[15:53:00] P1 Ep 17/80 | Train: 0.056255 | Val: 0.058606 | LR: 2.68e-04
|
| 36 |
+
[15:53:00] -> Saved P1 (val=0.058606)
|
| 37 |
+
[15:53:22] P1 Ep 18/80 | Train: 0.055517 | Val: 0.058299 | LR: 2.64e-04
|
| 38 |
+
[15:53:22] -> Saved P1 (val=0.058299)
|
| 39 |
+
[15:53:46] P1 Ep 19/80 | Train: 0.054703 | Val: 0.056898 | LR: 2.60e-04
|
| 40 |
+
[15:53:46] -> Saved P1 (val=0.056898)
|
| 41 |
+
[15:54:15] P1 Ep 20/80 | Train: 0.053777 | Val: 0.057646 | LR: 2.56e-04
|
| 42 |
+
[15:55:04] P1 Ep 21/80 | Train: 0.053153 | Val: 0.056199 | LR: 2.52e-04
|
| 43 |
+
[15:55:04] -> Saved P1 (val=0.056199)
|
| 44 |
+
[15:55:41] P1 Ep 22/80 | Train: 0.052464 | Val: 0.055376 | LR: 2.48e-04
|
| 45 |
+
[15:55:41] -> Saved P1 (val=0.055376)
|
| 46 |
+
[15:56:18] P1 Ep 23/80 | Train: 0.051747 | Val: 0.055485 | LR: 2.43e-04
|
| 47 |
+
[15:56:50] P1 Ep 24/80 | Train: 0.050900 | Val: 0.055214 | LR: 2.38e-04
|
| 48 |
+
[15:56:50] -> Saved P1 (val=0.055214)
|
| 49 |
+
[15:57:14] P1 Ep 25/80 | Train: 0.050435 | Val: 0.054996 | LR: 2.34e-04
|
| 50 |
+
[15:57:14] -> Saved P1 (val=0.054996)
|
| 51 |
+
[15:57:43] P1 Ep 26/80 | Train: 0.049864 | Val: 0.055223 | LR: 2.29e-04
|
| 52 |
+
[15:58:33] P1 Ep 27/80 | Train: 0.049200 | Val: 0.054005 | LR: 2.24e-04
|
| 53 |
+
[15:58:33] -> Saved P1 (val=0.054005)
|
| 54 |
+
[15:59:19] P1 Ep 28/80 | Train: 0.048483 | Val: 0.053483 | LR: 2.18e-04
|
| 55 |
+
[15:59:19] -> Saved P1 (val=0.053483)
|
| 56 |
+
[16:00:05] P1 Ep 29/80 | Train: 0.047984 | Val: 0.052937 | LR: 2.13e-04
|
| 57 |
+
[16:00:05] -> Saved P1 (val=0.052937)
|
| 58 |
+
[16:00:39] P1 Ep 30/80 | Train: 0.047200 | Val: 0.052694 | LR: 2.08e-04
|
| 59 |
+
[16:00:39] -> Saved P1 (val=0.052694)
|
| 60 |
+
[16:01:11] P1 Ep 31/80 | Train: 0.046624 | Val: 0.053359 | LR: 2.02e-04
|
| 61 |
+
[16:02:07] P1 Ep 32/80 | Train: 0.045999 | Val: 0.052335 | LR: 1.97e-04
|
| 62 |
+
[16:02:07] -> Saved P1 (val=0.052335)
|
| 63 |
+
[16:03:05] P1 Ep 33/80 | Train: 0.045480 | Val: 0.052796 | LR: 1.91e-04
|
| 64 |
+
[16:03:58] P1 Ep 34/80 | Train: 0.044809 | Val: 0.052662 | LR: 1.85e-04
|
| 65 |
+
[16:04:27] P1 Ep 35/80 | Train: 0.044315 | Val: 0.051852 | LR: 1.80e-04
|
| 66 |
+
[16:04:27] -> Saved P1 (val=0.051852)
|
| 67 |
+
[16:05:16] P1 Ep 36/80 | Train: 0.043599 | Val: 0.051811 | LR: 1.74e-04
|
| 68 |
+
[16:05:16] -> Saved P1 (val=0.051811)
|
| 69 |
+
[16:06:03] P1 Ep 37/80 | Train: 0.043138 | Val: 0.051202 | LR: 1.68e-04
|
| 70 |
+
[16:06:03] -> Saved P1 (val=0.051202)
|
| 71 |
+
[16:06:41] P1 Ep 38/80 | Train: 0.042539 | Val: 0.050464 | LR: 1.62e-04
|
| 72 |
+
[16:06:41] -> Saved P1 (val=0.050464)
|
| 73 |
+
[16:07:38] P1 Ep 39/80 | Train: 0.042066 | Val: 0.051059 | LR: 1.56e-04
|
| 74 |
+
[16:08:34] P1 Ep 40/80 | Train: 0.041315 | Val: 0.050918 | LR: 1.51e-04
|
| 75 |
+
[16:09:27] P1 Ep 41/80 | Train: 0.040930 | Val: 0.050821 | LR: 1.45e-04
|
| 76 |
+
[16:10:15] P1 Ep 42/80 | Train: 0.040304 | Val: 0.050680 | LR: 1.39e-04
|
| 77 |
+
[16:11:10] P1 Ep 43/80 | Train: 0.039895 | Val: 0.050563 | LR: 1.33e-04
|
| 78 |
+
[16:12:05] P1 Ep 44/80 | Train: 0.039291 | Val: 0.050782 | LR: 1.27e-04
|
| 79 |
+
[16:13:00] P1 Ep 45/80 | Train: 0.038875 | Val: 0.050635 | LR: 1.21e-04
|
| 80 |
+
[16:13:54] P1 Ep 46/80 | Train: 0.038366 | Val: 0.050127 | LR: 1.16e-04
|
| 81 |
+
[16:13:54] -> Saved P1 (val=0.050127)
|
| 82 |
+
[16:14:29] P1 Ep 47/80 | Train: 0.037889 | Val: 0.050041 | LR: 1.10e-04
|
| 83 |
+
[16:14:29] -> Saved P1 (val=0.050041)
|
| 84 |
+
[16:15:23] P1 Ep 48/80 | Train: 0.037326 | Val: 0.050060 | LR: 1.04e-04
|
| 85 |
+
[16:16:12] P1 Ep 49/80 | Train: 0.036947 | Val: 0.050365 | LR: 9.88e-05
|
| 86 |
+
[16:17:00] P1 Ep 50/80 | Train: 0.036355 | Val: 0.049527 | LR: 9.33e-05
|
| 87 |
+
[16:17:00] -> Saved P1 (val=0.049527)
|
| 88 |
+
[16:17:57] P1 Ep 51/80 | Train: 0.036024 | Val: 0.049895 | LR: 8.79e-05
|
| 89 |
+
[16:18:50] P1 Ep 52/80 | Train: 0.035437 | Val: 0.049741 | LR: 8.26e-05
|
| 90 |
+
[16:19:35] P1 Ep 53/80 | Train: 0.034996 | Val: 0.049694 | LR: 7.75e-05
|
| 91 |
+
[16:20:19] P1 Ep 54/80 | Train: 0.034717 | Val: 0.049697 | LR: 7.24e-05
|
| 92 |
+
[16:21:00] P1 Ep 55/80 | Train: 0.034225 | Val: 0.049615 | LR: 6.74e-05
|
| 93 |
+
[16:21:55] P1 Ep 56/80 | Train: 0.033776 | Val: 0.049514 | LR: 6.26e-05
|
| 94 |
+
[16:21:55] -> Saved P1 (val=0.049514)
|
| 95 |
+
[16:22:42] P1 Ep 57/80 | Train: 0.033460 | Val: 0.049664 | LR: 5.79e-05
|
| 96 |
+
[16:23:34] P1 Ep 58/80 | Train: 0.033085 | Val: 0.049812 | LR: 5.34e-05
|
| 97 |
+
[16:24:24] P1 Ep 59/80 | Train: 0.032784 | Val: 0.049797 | LR: 4.90e-05
|
| 98 |
+
[16:24:49] P1 Ep 60/80 | Train: 0.032414 | Val: 0.049671 | LR: 4.48e-05
|
| 99 |
+
[16:25:33] P1 Ep 61/80 | Train: 0.032062 | Val: 0.049591 | LR: 4.07e-05
|
| 100 |
+
[16:26:25] P1 Ep 62/80 | Train: 0.031821 | Val: 0.050111 | LR: 3.68e-05
|
| 101 |
+
[16:27:10] P1 Ep 63/80 | Train: 0.031511 | Val: 0.049890 | LR: 3.31e-05
|
| 102 |
+
[16:27:55] P1 Ep 64/80 | Train: 0.031259 | Val: 0.050195 | LR: 2.96e-05
|
| 103 |
+
[16:28:34] P1 Ep 65/80 | Train: 0.031016 | Val: 0.049906 | LR: 2.62e-05
|
| 104 |
+
[16:29:32] P1 Ep 66/80 | Train: 0.030782 | Val: 0.049883 | LR: 2.30e-05
|
| 105 |
+
[16:30:27] P1 Ep 67/80 | Train: 0.030587 | Val: 0.049873 | LR: 2.01e-05
|
| 106 |
+
[16:31:13] P1 Ep 68/80 | Train: 0.030401 | Val: 0.049917 | LR: 1.73e-05
|
| 107 |
+
[16:31:40] P1 Ep 69/80 | Train: 0.030268 | Val: 0.049848 | LR: 1.47e-05
|
| 108 |
+
[16:32:31] P1 Ep 70/80 | Train: 0.030057 | Val: 0.050109 | LR: 1.24e-05
|
| 109 |
+
[16:33:27] P1 Ep 71/80 | Train: 0.029958 | Val: 0.050045 | LR: 1.02e-05
|
| 110 |
+
[16:34:17] P1 Ep 72/80 | Train: 0.029819 | Val: 0.050060 | LR: 8.32e-06
|
| 111 |
+
[16:34:49] P1 Ep 73/80 | Train: 0.029710 | Val: 0.050139 | LR: 6.61e-06
|
| 112 |
+
[16:35:11] P1 Ep 74/80 | Train: 0.029600 | Val: 0.050150 | LR: 5.13e-06
|
| 113 |
+
[16:35:48] P1 Ep 75/80 | Train: 0.029548 | Val: 0.050126 | LR: 3.87e-06
|
| 114 |
+
[16:36:41] P1 Ep 76/80 | Train: 0.029521 | Val: 0.050131 | LR: 2.84e-06
|
| 115 |
+
[16:37:25] P1 Ep 77/80 | Train: 0.029491 | Val: 0.050194 | LR: 2.04e-06
|
| 116 |
+
[16:38:05] P1 Ep 78/80 | Train: 0.029450 | Val: 0.050196 | LR: 1.46e-06
|
| 117 |
+
[16:38:35] P1 Ep 79/80 | Train: 0.029430 | Val: 0.050184 | LR: 1.12e-06
|
| 118 |
+
[16:39:27] P1 Ep 80/80 | Train: 0.029404 | Val: 0.050189 | LR: 1.00e-06
|
| 119 |
+
[16:39:27] === Phase 2: 2-step AR (40 epochs, noise=0.0078) ===
|
| 120 |
+
[16:39:32] Train: 44034, Val: 5412
|
| 121 |
+
[16:40:26] P2 Ep 1/40 | Train: 0.047285 | Val: 0.070376 | LR: 9.98e-05
|
| 122 |
+
[16:40:26] -> Saved P2 (val=0.070376)
|
| 123 |
+
[16:41:17] P2 Ep 2/40 | Train: 0.046644 | Val: 0.069826 | LR: 9.94e-05
|
| 124 |
+
[16:41:17] -> Saved P2 (val=0.069826)
|
| 125 |
+
[16:42:04] P2 Ep 3/40 | Train: 0.046104 | Val: 0.069656 | LR: 9.86e-05
|
| 126 |
+
[16:42:04] -> Saved P2 (val=0.069656)
|
| 127 |
+
[16:43:01] P2 Ep 4/40 | Train: 0.045620 | Val: 0.070126 | LR: 9.76e-05
|
| 128 |
+
[16:43:58] P2 Ep 5/40 | Train: 0.045143 | Val: 0.069811 | LR: 9.62e-05
|
| 129 |
+
[16:44:42] P2 Ep 6/40 | Train: 0.044883 | Val: 0.069925 | LR: 9.46e-05
|
| 130 |
+
[16:45:26] P2 Ep 7/40 | Train: 0.044342 | Val: 0.070318 | LR: 9.27e-05
|
| 131 |
+
[16:46:14] P2 Ep 8/40 | Train: 0.043952 | Val: 0.070639 | LR: 9.05e-05
|
| 132 |
+
[16:46:59] P2 Ep 9/40 | Train: 0.043475 | Val: 0.070085 | LR: 8.81e-05
|
| 133 |
+
[16:47:45] P2 Ep 10/40 | Train: 0.043026 | Val: 0.069909 | LR: 8.55e-05
|
| 134 |
+
[16:48:27] P2 Ep 11/40 | Train: 0.042577 | Val: 0.070436 | LR: 8.26e-05
|
| 135 |
+
[16:49:15] P2 Ep 12/40 | Train: 0.042101 | Val: 0.071131 | LR: 7.96e-05
|
| 136 |
+
[16:50:02] P2 Ep 13/40 | Train: 0.041770 | Val: 0.070430 | LR: 7.64e-05
|
| 137 |
+
[16:50:47] P2 Ep 14/40 | Train: 0.041258 | Val: 0.070631 | LR: 7.30e-05
|
| 138 |
+
[16:51:29] P2 Ep 15/40 | Train: 0.040938 | Val: 0.071101 | LR: 6.94e-05
|
| 139 |
+
[16:52:15] P2 Ep 16/40 | Train: 0.040437 | Val: 0.070787 | LR: 6.58e-05
|
| 140 |
+
[16:52:59] P2 Ep 17/40 | Train: 0.040043 | Val: 0.071084 | LR: 6.21e-05
|
| 141 |
+
[16:53:41] P2 Ep 18/40 | Train: 0.039619 | Val: 0.070765 | LR: 5.82e-05
|
| 142 |
+
[16:54:24] P2 Ep 19/40 | Train: 0.039209 | Val: 0.070925 | LR: 5.44e-05
|
| 143 |
+
[16:55:05] P2 Ep 20/40 | Train: 0.038831 | Val: 0.070665 | LR: 5.05e-05
|
| 144 |
+
[16:55:48] P2 Ep 21/40 | Train: 0.038462 | Val: 0.071706 | LR: 4.66e-05
|
| 145 |
+
[16:56:29] P2 Ep 22/40 | Train: 0.038149 | Val: 0.070848 | LR: 4.28e-05
|
| 146 |
+
[16:57:10] P2 Ep 23/40 | Train: 0.037821 | Val: 0.071886 | LR: 3.89e-05
|
| 147 |
+
[16:57:48] P2 Ep 24/40 | Train: 0.037490 | Val: 0.071343 | LR: 3.52e-05
|
| 148 |
+
[16:58:28] P2 Ep 25/40 | Train: 0.037162 | Val: 0.071360 | LR: 3.16e-05
|
| 149 |
+
[16:59:07] P2 Ep 26/40 | Train: 0.036869 | Val: 0.071390 | LR: 2.80e-05
|
| 150 |
+
[16:59:46] P2 Ep 27/40 | Train: 0.036626 | Val: 0.071814 | LR: 2.46e-05
|
| 151 |
+
[17:00:25] P2 Ep 28/40 | Train: 0.036370 | Val: 0.071899 | LR: 2.14e-05
|
| 152 |
+
[17:01:08] P2 Ep 29/40 | Train: 0.036102 | Val: 0.072189 | LR: 1.84e-05
|
| 153 |
+
[17:01:50] P2 Ep 30/40 | Train: 0.035936 | Val: 0.071791 | LR: 1.55e-05
|
| 154 |
+
[17:02:28] P2 Ep 31/40 | Train: 0.035759 | Val: 0.072592 | LR: 1.29e-05
|
| 155 |
+
[17:03:07] P2 Ep 32/40 | Train: 0.035663 | Val: 0.072196 | LR: 1.05e-05
|
| 156 |
+
[17:04:04] P2 Ep 33/40 | Train: 0.035482 | Val: 0.072056 | LR: 8.29e-06
|
| 157 |
+
[17:04:50] P2 Ep 34/40 | Train: 0.035372 | Val: 0.072318 | LR: 6.40e-06
|
| 158 |
+
[17:05:37] P2 Ep 35/40 | Train: 0.035295 | Val: 0.072516 | LR: 4.77e-06
|
| 159 |
+
[17:06:23] P2 Ep 36/40 | Train: 0.035200 | Val: 0.072428 | LR: 3.42e-06
|
| 160 |
+
[17:07:03] P2 Ep 37/40 | Train: 0.035132 | Val: 0.072473 | LR: 2.37e-06
|
| 161 |
+
[17:08:03] P2 Ep 38/40 | Train: 0.035041 | Val: 0.072513 | LR: 1.61e-06
|
| 162 |
+
[17:08:52] P2 Ep 39/40 | Train: 0.035027 | Val: 0.072532 | LR: 1.15e-06
|
| 163 |
+
[17:09:36] P2 Ep 40/40 | Train: 0.035019 | Val: 0.072462 | LR: 1.00e-06
|
| 164 |
+
[17:09:36] === Phase 3: 4-step AR (40 epochs, noise=0.0039) ===
|
| 165 |
+
[17:09:41] Train: 10919, Val: 1342
|
| 166 |
+
[17:10:22] P3 Ep 1/40 | Train: 0.066989 | Val: 0.105359 | LR: 4.99e-05
|
| 167 |
+
[17:10:22] -> Saved P3 (val=0.105359)
|
| 168 |
+
[17:11:01] P3 Ep 2/40 | Train: 0.065233 | Val: 0.105719 | LR: 4.97e-05
|
| 169 |
+
[17:11:41] P3 Ep 3/40 | Train: 0.064332 | Val: 0.105288 | LR: 4.93e-05
|
| 170 |
+
[17:11:41] -> Saved P3 (val=0.105288)
|
| 171 |
+
[17:12:20] P3 Ep 4/40 | Train: 0.063536 | Val: 0.106669 | LR: 4.88e-05
|
| 172 |
+
[17:12:57] P3 Ep 5/40 | Train: 0.063039 | Val: 0.106552 | LR: 4.81e-05
|
| 173 |
+
[17:13:32] P3 Ep 6/40 | Train: 0.062253 | Val: 0.106537 | LR: 4.73e-05
|
| 174 |
+
[17:14:09] P3 Ep 7/40 | Train: 0.061609 | Val: 0.107229 | LR: 4.64e-05
|
| 175 |
+
[17:14:47] P3 Ep 8/40 | Train: 0.061302 | Val: 0.106256 | LR: 4.53e-05
|
| 176 |
+
[17:15:27] P3 Ep 9/40 | Train: 0.060679 | Val: 0.108110 | LR: 4.41e-05
|
| 177 |
+
[17:16:06] P3 Ep 10/40 | Train: 0.060154 | Val: 0.109534 | LR: 4.28e-05
|
| 178 |
+
[17:16:44] P3 Ep 11/40 | Train: 0.059723 | Val: 0.108979 | LR: 4.14e-05
|
| 179 |
+
[17:17:18] P3 Ep 12/40 | Train: 0.059178 | Val: 0.107284 | LR: 3.99e-05
|
| 180 |
+
[17:17:51] P3 Ep 13/40 | Train: 0.058841 | Val: 0.108774 | LR: 3.83e-05
|
| 181 |
+
[17:18:25] P3 Ep 14/40 | Train: 0.058297 | Val: 0.109313 | LR: 3.66e-05
|
| 182 |
+
[17:18:59] P3 Ep 15/40 | Train: 0.057755 | Val: 0.108477 | LR: 3.49e-05
|
| 183 |
+
[17:19:35] P3 Ep 16/40 | Train: 0.057522 | Val: 0.108420 | LR: 3.31e-05
|
| 184 |
+
[17:20:11] P3 Ep 17/40 | Train: 0.057237 | Val: 0.110192 | LR: 3.12e-05
|
| 185 |
+
[17:20:48] P3 Ep 18/40 | Train: 0.056578 | Val: 0.108483 | LR: 2.93e-05
|
| 186 |
+
[17:21:27] P3 Ep 19/40 | Train: 0.056250 | Val: 0.108837 | LR: 2.74e-05
|
| 187 |
+
[17:22:06] P3 Ep 20/40 | Train: 0.055933 | Val: 0.109871 | LR: 2.55e-05
|
| 188 |
+
[17:22:44] P3 Ep 21/40 | Train: 0.055610 | Val: 0.109395 | LR: 2.36e-05
|
| 189 |
+
[17:23:21] P3 Ep 22/40 | Train: 0.055233 | Val: 0.109832 | LR: 2.17e-05
|
| 190 |
+
[17:23:58] P3 Ep 23/40 | Train: 0.054883 | Val: 0.110014 | LR: 1.98e-05
|
| 191 |
+
[17:24:36] P3 Ep 24/40 | Train: 0.054613 | Val: 0.109150 | LR: 1.79e-05
|
| 192 |
+
[17:25:15] P3 Ep 25/40 | Train: 0.054338 | Val: 0.110191 | LR: 1.61e-05
|
| 193 |
+
[17:25:54] P3 Ep 26/40 | Train: 0.054093 | Val: 0.109644 | LR: 1.44e-05
|
| 194 |
+
[17:26:34] P3 Ep 27/40 | Train: 0.053798 | Val: 0.110244 | LR: 1.27e-05
|
| 195 |
+
[17:27:13] P3 Ep 28/40 | Train: 0.053642 | Val: 0.110176 | LR: 1.11e-05
|
| 196 |
+
[17:27:53] P3 Ep 29/40 | Train: 0.053352 | Val: 0.110111 | LR: 9.59e-06
|
| 197 |
+
[17:28:33] P3 Ep 30/40 | Train: 0.053254 | Val: 0.110057 | LR: 8.18e-06
|
| 198 |
+
[17:29:13] P3 Ep 31/40 | Train: 0.053071 | Val: 0.110303 | LR: 6.87e-06
|
| 199 |
+
[17:29:50] P3 Ep 32/40 | Train: 0.052807 | Val: 0.110612 | LR: 5.68e-06
|
| 200 |
+
[17:30:25] P3 Ep 33/40 | Train: 0.052767 | Val: 0.110462 | LR: 4.61e-06
|
| 201 |
+
[17:31:03] P3 Ep 34/40 | Train: 0.052605 | Val: 0.110954 | LR: 3.67e-06
|
| 202 |
+
[17:31:41] P3 Ep 35/40 | Train: 0.052573 | Val: 0.110939 | LR: 2.86e-06
|
| 203 |
+
[17:32:19] P3 Ep 36/40 | Train: 0.052437 | Val: 0.110889 | LR: 2.20e-06
|
| 204 |
+
[17:32:57] P3 Ep 37/40 | Train: 0.052463 | Val: 0.111071 | LR: 1.68e-06
|
| 205 |
+
[17:33:35] P3 Ep 38/40 | Train: 0.052356 | Val: 0.110856 | LR: 1.30e-06
|
| 206 |
+
[17:34:13] P3 Ep 39/40 | Train: 0.052375 | Val: 0.111090 | LR: 1.08e-06
|
| 207 |
+
[17:34:52] P3 Ep 40/40 | Train: 0.052406 | Val: 0.111096 | LR: 1.00e-06
|
| 208 |
+
[17:34:52] === Phase 4: 8-step AR (40 epochs, noise=0.0000) ===
|
| 209 |
+
[17:34:57] Train: 2685, Val: 330
|
| 210 |
+
[17:35:27] P4 Ep 1/40 | Train: 0.101111 | Val: 0.162571 | LR: 2.00e-05
|
| 211 |
+
[17:35:27] -> Saved P4 (val=0.162571)
|
| 212 |
+
[17:35:57] P4 Ep 2/40 | Train: 0.098553 | Val: 0.160688 | LR: 1.99e-05
|
| 213 |
+
[17:35:58] -> Saved P4 (val=0.160688)
|
| 214 |
+
[17:36:27] P4 Ep 3/40 | Train: 0.097715 | Val: 0.159298 | LR: 1.97e-05
|
| 215 |
+
[17:36:27] -> Saved P4 (val=0.159298)
|
| 216 |
+
[17:36:57] P4 Ep 4/40 | Train: 0.096368 | Val: 0.159610 | LR: 1.95e-05
|
| 217 |
+
[17:37:27] P4 Ep 5/40 | Train: 0.095401 | Val: 0.159243 | LR: 1.93e-05
|
| 218 |
+
[17:37:27] -> Saved P4 (val=0.159243)
|
| 219 |
+
[17:37:58] P4 Ep 6/40 | Train: 0.094479 | Val: 0.160074 | LR: 1.90e-05
|
| 220 |
+
[17:38:27] P4 Ep 7/40 | Train: 0.093345 | Val: 0.160185 | LR: 1.86e-05
|
| 221 |
+
[17:38:57] P4 Ep 8/40 | Train: 0.092778 | Val: 0.163288 | LR: 1.82e-05
|
| 222 |
+
[17:39:27] P4 Ep 9/40 | Train: 0.091566 | Val: 0.163134 | LR: 1.77e-05
|
| 223 |
+
[17:39:57] P4 Ep 10/40 | Train: 0.091514 | Val: 0.161680 | LR: 1.72e-05
|
| 224 |
+
[17:40:27] P4 Ep 11/40 | Train: 0.090770 | Val: 0.162345 | LR: 1.67e-05
|
| 225 |
+
[17:40:57] P4 Ep 12/40 | Train: 0.090370 | Val: 0.163773 | LR: 1.61e-05
|
| 226 |
+
[17:41:27] P4 Ep 13/40 | Train: 0.089234 | Val: 0.161309 | LR: 1.55e-05
|
| 227 |
+
[17:41:57] P4 Ep 14/40 | Train: 0.088393 | Val: 0.164177 | LR: 1.48e-05
|
| 228 |
+
[17:42:27] P4 Ep 15/40 | Train: 0.087783 | Val: 0.161810 | LR: 1.41e-05
|
| 229 |
+
[17:42:57] P4 Ep 16/40 | Train: 0.087447 | Val: 0.164077 | LR: 1.34e-05
|
| 230 |
+
[17:43:26] P4 Ep 17/40 | Train: 0.087248 | Val: 0.162098 | LR: 1.27e-05
|
| 231 |
+
[17:43:56] P4 Ep 18/40 | Train: 0.086734 | Val: 0.163517 | LR: 1.20e-05
|
| 232 |
+
[17:44:26] P4 Ep 19/40 | Train: 0.086703 | Val: 0.163718 | LR: 1.12e-05
|
| 233 |
+
[17:44:56] P4 Ep 20/40 | Train: 0.085615 | Val: 0.162694 | LR: 1.05e-05
|
| 234 |
+
[17:45:26] P4 Ep 21/40 | Train: 0.085481 | Val: 0.163687 | LR: 9.75e-06
|
| 235 |
+
[17:45:55] P4 Ep 22/40 | Train: 0.085017 | Val: 0.164132 | LR: 9.01e-06
|
| 236 |
+
[17:46:25] P4 Ep 23/40 | Train: 0.084124 | Val: 0.164077 | LR: 8.28e-06
|
| 237 |
+
[17:46:56] P4 Ep 24/40 | Train: 0.084426 | Val: 0.164232 | LR: 7.56e-06
|
| 238 |
+
[17:47:26] P4 Ep 25/40 | Train: 0.084260 | Val: 0.164044 | LR: 6.86e-06
|
| 239 |
+
[17:47:56] P4 Ep 26/40 | Train: 0.083558 | Val: 0.164989 | LR: 6.19e-06
|
| 240 |
+
[17:48:26] P4 Ep 27/40 | Train: 0.083491 | Val: 0.165989 | LR: 5.54e-06
|
| 241 |
+
[17:48:56] P4 Ep 28/40 | Train: 0.082750 | Val: 0.165438 | LR: 4.92e-06
|
| 242 |
+
[17:49:26] P4 Ep 29/40 | Train: 0.082852 | Val: 0.165827 | LR: 4.33e-06
|
| 243 |
+
[17:49:56] P4 Ep 30/40 | Train: 0.082666 | Val: 0.166003 | LR: 3.78e-06
|
| 244 |
+
[17:50:26] P4 Ep 31/40 | Train: 0.082712 | Val: 0.165605 | LR: 3.28e-06
|
| 245 |
+
[17:50:56] P4 Ep 32/40 | Train: 0.082512 | Val: 0.166602 | LR: 2.81e-06
|
| 246 |
+
[17:51:26] P4 Ep 33/40 | Train: 0.081745 | Val: 0.166517 | LR: 2.40e-06
|
| 247 |
+
[17:51:55] P4 Ep 34/40 | Train: 0.082233 | Val: 0.166963 | LR: 2.04e-06
|
| 248 |
+
[17:52:25] P4 Ep 35/40 | Train: 0.081910 | Val: 0.167685 | LR: 1.72e-06
|
| 249 |
+
[17:52:55] P4 Ep 36/40 | Train: 0.081687 | Val: 0.167599 | LR: 1.46e-06
|
| 250 |
+
[17:53:33] P4 Ep 37/40 | Train: 0.081693 | Val: 0.166911 | LR: 1.26e-06
|
| 251 |
+
[17:54:05] P4 Ep 38/40 | Train: 0.081608 | Val: 0.167503 | LR: 1.12e-06
|
| 252 |
+
[17:54:35] P4 Ep 39/40 | Train: 0.081655 | Val: 0.167355 | LR: 1.03e-06
|
| 253 |
+
[17:55:04] P4 Ep 40/40 | Train: 0.081653 | Val: 0.167608 | LR: 1.00e-06
|
| 254 |
+
[17:55:04] Training complete. Now score each phase to find best.
|