guychuk commited on
Commit
6caa701
Β·
verified Β·
1 Parent(s): b2041f7

v8: model card with LOAO results + plots

Browse files
Files changed (1) hide show
  1. README.md +33 -8
README.md CHANGED
@@ -14,23 +14,48 @@ language:
14
  library_name: pytorch
15
  ---
16
 
17
- # Flight-JEPA v7
18
 
19
  A trajectory forecasting model for aircraft on terminal-area approach,
20
  specialized for the **blindspot continuation** task: given an observed
21
  past track, predict the trajectory through a coverage gap of variable
22
  length and the reappearance distribution.
23
 
24
- The headline contribution of v7 is a **JEPA-style past-track masked
25
  pretraining recipe** that produces representations more robust to test-
26
- time radar coverage gaps. Pretrained-then-fine-tuned models maintain
27
- significantly lower forecasting error and better calibrated uncertainty
28
- when up to half of the past observations are missing β€” the regime
29
- aviation deployment cares about.
 
30
 
31
- ## Quick numbers
32
 
33
- On RKSIa (Incheon arrivals, 8092 test trajectories, n=3 seeds):
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
 
35
  | Past-track dropout | Scratch FDE | Pretrained FDE | Ξ” | p (Welch) |
36
  |---|---:|---:|---:|---:|
 
14
  library_name: pytorch
15
  ---
16
 
17
+ # Flight-JEPA v8
18
 
19
  A trajectory forecasting model for aircraft on terminal-area approach,
20
  specialized for the **blindspot continuation** task: given an observed
21
  past track, predict the trajectory through a coverage gap of variable
22
  length and the reappearance distribution.
23
 
24
+ The headline contribution is a **JEPA-style past-track masked
25
  pretraining recipe** that produces representations more robust to test-
26
+ time radar coverage gaps, with the gain *generalizing across airports*.
27
+ Pretrained-then-fine-tuned models maintain significantly lower
28
+ forecasting error and better-calibrated uncertainty when up to 70% of
29
+ past observations are missing β€” including on **completely held-out
30
+ airports the fine-tuning never saw**.
31
 
32
+ ## v8 β€” leave-one-airport-out (LOAO) headline
33
 
34
+ Across 4 LOAO folds (held out: RKSIa / RKSId / ESSA / LSZH, n=3 seeds
35
+ each), pretrained beats scratch with significance:
36
+
37
+ | Past-track dropout | Mean Ξ” FDE | p |
38
+ |---|---:|---:|
39
+ | 0% (clean β€” no regression) | +2.8% | 0.41 |
40
+ | 30% | βˆ’6.6% | 0.04 βœ“ |
41
+ | **50%** | **βˆ’23.4%** | **<0.001 βœ“** |
42
+ | **70%** | **βˆ’22.5%** | **<0.001 βœ“** |
43
+
44
+ 11 of 12 comparisons at β‰₯30% dropout reach p<0.05. All 4 LOAO folds
45
+ pass the locked criterion independently.
46
+
47
+ ![v8 summary](plots/summary_v8.png)
48
+ ![v8 FDE per airport](plots/fde_per_airport.png)
49
+ ![v8 coverage per airport](plots/coverage_per_airport.png)
50
+
51
+ The result generalizes across very different airports (Korea, Sweden,
52
+ Switzerland β€” different runway geometries and procedures). It uses an
53
+ airport-ID token (UniTraj recipe, arxiv:2403.15098) for conditioning.
54
+
55
+ ## v7 β€” single-airport reference
56
+
57
+ The v7 prerequisite (single-airport, RKSIa-only) showed the same effect
58
+ on a within-airport split:
59
 
60
  | Past-track dropout | Scratch FDE | Pretrained FDE | Ξ” | p (Welch) |
61
  |---|---:|---:|---:|---:|