ethan1115 commited on
Commit
0161e74
·
verified ·
1 Parent(s): 8c0d77a

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +19 -0
  2. transfer/code/CCFM/CCFM_Architecture.md +437 -0
  3. transfer/code/CCFM/_bootstrap_scdfm.py +101 -0
  4. transfer/code/CCFM/cell-eval-outdir/agg_results.csv +10 -0
  5. transfer/code/CCFM/cell-eval-outdir/pred_de.csv +0 -0
  6. transfer/code/CCFM/cell-eval-outdir/real_de.csv +0 -0
  7. transfer/code/CCFM/cell-eval-outdir/results.csv +40 -0
  8. transfer/code/CCFM/config/config_cascaded.py +92 -0
  9. transfer/code/CCFM/config/config_cascaded.py.bak +89 -0
  10. transfer/code/CCFM/eval_ccfm_v2.sh +68 -0
  11. transfer/code/CCFM/eval_joint_generate.sh +56 -0
  12. transfer/code/CCFM/eval_joint_generate.sh.5404217.stats +467 -0
  13. transfer/code/CCFM/logs/ccfm_1gpu_cached_5404438.out +57 -0
  14. transfer/code/CCFM/logs/ccfm_1gpu_cached_5404526.out +68 -0
  15. transfer/code/CCFM/logs/ccfm_1gpu_online_5404417.out +3 -0
  16. transfer/code/CCFM/logs/ccfm_topk30_neg_5402619.out +111 -0
  17. transfer/code/CCFM/logs/ccfm_v2_5404727.out +214 -0
  18. transfer/code/CCFM/logs/ccfm_v2_5404735.out +3 -0
  19. transfer/code/CCFM/logs/ccfm_v2_cached_5404728.out +236 -0
  20. transfer/code/CCFM/logs/ccfm_v2_cached_5404736.out +236 -0
  21. transfer/code/CCFM/logs/ccfm_v2_cached_5404737.out +236 -0
  22. transfer/code/CCFM/logs/ccfm_v2_cached_5405774.out +0 -0
  23. transfer/code/CCFM/logs/ccfm_v2_resume_5406605.out +3 -0
  24. transfer/code/CCFM/logs/eval_ccfm_v2_5406214.out +437 -0
  25. transfer/code/CCFM/logs/preextract_5402629.out +1 -0
  26. transfer/code/CCFM/logs/preextract_5402631.out +36 -0
  27. transfer/code/CCFM/plan.md +490 -0
  28. transfer/code/CCFM/preextract_scgpt.5402631.stats +467 -0
  29. transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/agg_results.csv +10 -0
  30. transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/checkpoint.pt +3 -0
  31. transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/pred.h5ad +3 -0
  32. transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/real.h5ad +3 -0
  33. transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/results.csv +40 -0
  34. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/iteration_10000/checkpoint.pt +3 -0
  35. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/loss_curve.csv +0 -0
  36. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/tb_logs/events.out.tfevents.1773518027.b0021.201.0 +3 -0
  37. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/tb_logs/events.out.tfevents.1773518372.b0021.200.0 +3 -0
  38. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/tb_logs/events.out.tfevents.1773518599.b0021.200.0 +3 -0
  39. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/tb_logs/events.out.tfevents.1773561648.b0034.161.0 +3 -0
  40. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/eval_only/agg_results.csv +10 -0
  41. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/eval_only/pred.h5ad +3 -0
  42. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/eval_only/real.h5ad +3 -0
  43. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/eval_only/results.csv +40 -0
  44. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/agg_results.csv +10 -0
  45. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/checkpoint.pt +3 -0
  46. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/pred.h5ad +3 -0
  47. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/real.h5ad +3 -0
  48. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/results.csv +40 -0
  49. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/loss_curve.csv +3 -0
  50. transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/tb_logs/events.out.tfevents.1773517868.b0019.200.0 +3 -0
.gitattributes CHANGED
@@ -58,3 +58,22 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
58
  # Video files - compressed
59
  *.mp4 filter=lfs diff=lfs merge=lfs -text
60
  *.webm filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
58
  # Video files - compressed
59
  *.mp4 filter=lfs diff=lfs merge=lfs -text
60
  *.webm filter=lfs diff=lfs merge=lfs -text
61
+ transfer/code/CCFM/logs/ccfm_1gpu_online_5404417.out filter=lfs diff=lfs merge=lfs -text
62
+ transfer/code/CCFM/logs/ccfm_v2_5404735.out filter=lfs diff=lfs merge=lfs -text
63
+ transfer/code/CCFM/logs/ccfm_v2_resume_5406605.out filter=lfs diff=lfs merge=lfs -text
64
+ transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/pred.h5ad filter=lfs diff=lfs merge=lfs -text
65
+ transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/real.h5ad filter=lfs diff=lfs merge=lfs -text
66
+ transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/eval_only/pred.h5ad filter=lfs diff=lfs merge=lfs -text
67
+ transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/eval_only/real.h5ad filter=lfs diff=lfs merge=lfs -text
68
+ transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/pred.h5ad filter=lfs diff=lfs merge=lfs -text
69
+ transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/real.h5ad filter=lfs diff=lfs merge=lfs -text
70
+ transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/loss_curve.csv filter=lfs diff=lfs merge=lfs -text
71
+ transfer/code/CCFM/result_joint_generate/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-online/eval_only/pred.h5ad filter=lfs diff=lfs merge=lfs -text
72
+ transfer/code/CCFM/result_joint_generate/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-online/eval_only/real.h5ad filter=lfs diff=lfs merge=lfs -text
73
+ transfer/code/CCFM/test_result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_0/pred.h5ad filter=lfs diff=lfs merge=lfs -text
74
+ transfer/code/CCFM/test_result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_0/real.h5ad filter=lfs diff=lfs merge=lfs -text
75
+ transfer/code/ori_scDFM/data/norman.h5ad filter=lfs diff=lfs merge=lfs -text
76
+ transfer/code/ori_scDFM/logs/ccfm_baseline_5404238.out filter=lfs diff=lfs merge=lfs -text
77
+ transfer/code/scDFM/data/norman/go.csv filter=lfs diff=lfs merge=lfs -text
78
+ transfer/code/scDFM/data/norman/perturb_processed.h5ad filter=lfs diff=lfs merge=lfs -text
79
+ transfer/code/scDFM/data/norman.h5ad filter=lfs diff=lfs merge=lfs -text
transfer/code/CCFM/CCFM_Architecture.md ADDED
@@ -0,0 +1,437 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # CCFM (Cascaded Conditioned Flow Matching) 架构文档
2
+
3
+ ## 一、项目概述
4
+
5
+ CCFM 是一个**级联流匹配**框架,融合三个模型的优势来做单细胞扰动预测:
6
+
7
+ - **scDFM**:基础流匹配架构(backbone、数据加载、训练范式)
8
+ - **LatentForcing**:级联双流思想(latent 流 + 表达流分阶段生成)
9
+ - **scGPT**:冻结的预训练模型,提供 per-gene 上下文化特征
10
+
11
+ ### 核心创新
12
+
13
+ 借鉴 LatentForcing 的双流架构,将其从图像域迁移到单细胞域:
14
+
15
+ | LatentForcing (图像) | CCFM (单细胞) |
16
+ |---|---|
17
+ | 像素值 | 基因表达值 |
18
+ | DINO-v2 特征(辅助生成目标) | scGPT 上下文特征(辅助生成目标) |
19
+ | 类别标签(条件信号) | control 表达 + perturbation_id(条件信号) |
20
+
21
+ **关键区分**:scGPT 特征是从 target(扰动细胞)提取的**辅助生成目标**,不是条件信号。推理时模型从噪声生成 scGPT 特征(Stage 1),再用生成的特征引导表达值生成(Stage 2)。真正的条件信号是 control 表达和 perturbation_id,它们在推理时始终可获取。
22
+
23
+ ### 文件结构
24
+
25
+ ```
26
+ CCFM/
27
+ ├── _bootstrap_scdfm.py # Bootstrap scDFM 模块,命名空间隔离
28
+ ├── config/
29
+ │ └── config_cascaded.py # CascadedFlowConfig dataclass (tyro CLI)
30
+ ├── src/
31
+ │ ├── __init__.py
32
+ │ ├── utils.py # Re-exports scDFM utils
33
+ │ ├── _scdfm_imports.py # scDFM 模块集中导入
34
+ │ ├── denoiser.py # CascadedDenoiser (训练/推理逻辑)
35
+ │ ├── model/
36
+ │ │ ├── model.py # CascadedFlowModel 双流模型
37
+ │ │ └── layers.py # LatentEmbedder, LatentDecoder
38
+ │ └── data/
39
+ │ ├── data.py # scDFM 数据加载集成
40
+ │ └── scgpt_extractor.py # FrozenScGPTExtractor
41
+ ├── scripts/
42
+ │ ├── run_cascaded.py # 训练/评估入口
43
+ │ └── download_scgpt.py # 下载 scGPT 预训练模型
44
+ ├── run.sh # pjsub 模板
45
+ └── run_topk30_neg.sh # 完整参数化 job 脚本
46
+ ```
47
+
48
+ ### 默认超参数
49
+
50
+ | 参数 | 值 | 说明 |
51
+ |---|---|---|
52
+ | `B` | 48 | batch size |
53
+ | `G_full` | 5000 | HVG 总数 |
54
+ | `G` | 1000 | 训练时随机基因子集 |
55
+ | `d_model` | 128 | 隐藏维度 |
56
+ | `scgpt_dim` | 512 | scGPT 输出特征维度 |
57
+ | `bottleneck_dim` | 128 | LatentEmbedder 瓶颈维度 |
58
+ | `nhead` | 8 | 注意力头数 |
59
+ | `nlayers` | 4 | Transformer 层数 |
60
+ | `dh_depth` | 2 | LatentDecoder block 数 |
61
+ | `choose_latent_p` | 0.4 | 训练 latent 流的概率 |
62
+ | `latent_weight` | 1.0 | latent loss 权重 |
63
+ | `gamma` | 0.5 | MMD loss 权重 |
64
+ | `lr` | 5e-5 | 学习率 |
65
+ | `steps` | 200000 | 训练总步数 |
66
+ | `latent_steps` | 20 | 推理 ODE 步数(latent) |
67
+ | `expr_steps` | 20 | 推理 ODE 步数(表达) |
68
+ | `warmup_batches` | 200 | scGPT running stats 预热 |
69
+
70
+ ---
71
+
72
+ ## 二、训练流程 Tensor 数据流
73
+
74
+ ### 2.1 数据准备
75
+
76
+ ```
77
+ 输入:
78
+ source: (B, G_full=5000) 控制组表达
79
+ target: (B, G_full=5000) 扰动后表达
80
+ perturbation_id: (B, 2) 扰动 token ID
81
+ gene_ids: (G_full=5000,) 全部基因的 vocab ID
82
+
83
+ 随机采样 1000 个基因:
84
+ input_gene_ids = randperm(5000)[:1000] → (1000,)
85
+ source_sub = source[:, input_gene_ids] → (48, 1000)
86
+ target_sub = target[:, input_gene_ids] → (48, 1000)
87
+ gene_input = gene_ids[input_gene_ids].expand → (48, 1000)
88
+ ```
89
+
90
+ ### 2.2 冻结 scGPT 提取辅助生成目标(类似 LatentForcing 的 DINO 特征提取)
91
+
92
+ ```
93
+ scgpt_extractor.extract(target_sub, gene_indices=input_gene_ids):
94
+
95
+ target_sub: (48, 1000) 输入表达
96
+ hvg_ids: (1000,) HVG → scGPT vocab ID 映射
97
+ valid_mask: (1000,) bool, 过滤在 scGPT vocab 中的基因
98
+ expr_valid: (48, G_valid) 有效基因的表达值
99
+
100
+ 若 G_valid+1 > max_seq_len(1200):
101
+ 随机采样 1199 个基因
102
+ seq_len = 1200
103
+ 否则:
104
+ seq_len = G_valid + 1
105
+
106
+ 拼接 CLS token:
107
+ src = [cls_id | gene_ids] → (48, seq_len) long
108
+ values = [0 | expr_valid] → (48, seq_len) float
109
+
110
+ scGPT frozen forward:
111
+ encoder_out = scgpt._encode(src, values, mask) → (48, seq_len, 512)
112
+
113
+ 去掉 CLS, scatter 回固定位置:
114
+ gene_features = encoder_out[:, 1:, :] → (48, seq_len-1, 512)
115
+ output = zeros(48, 1000, 512)
116
+ output.scatter_(1, idx, gene_features) → (48, 1000, 512)
117
+
118
+ 归一化 (running mean/var, warmup 200 batches 后冻结):
119
+ output = (output - running_mean) / sqrt(running_var) * target_std
120
+
121
+ z_target: (48, 1000, 512)
122
+ ```
123
+
124
+ ### 2.3 级联时间采样
125
+
126
+ ```
127
+ t_latent = rand(48) → (48,)
128
+ t_expr = rand(48) → (48,)
129
+ choose_latent_mask = rand(48) < 0.4 → (48,) bool
130
+
131
+ 对每个样本二选一:
132
+ 若 mask=True (40%概率): t_latent 保留, t_expr=0, w_expr=0, w_latent=1 → 训练 latent 流
133
+ 若 mask=False (60%概率): t_expr 保留, t_latent=1, w_expr=1, w_latent=0 → 训练表达流
134
+
135
+ 输出:
136
+ t_expr: (48,) 表达流时间步
137
+ t_latent: (48,) 潜变量流时间步
138
+ w_expr: (48,) 表达 loss 权重 (0 或 1)
139
+ w_latent: (48,) 潜变量 loss 权重 (0 或 1)
140
+ ```
141
+
142
+ ### 2.4 Flow Path 采样(线性插值 + 速度)
143
+
144
+ ```
145
+ 表达流:
146
+ noise_expr = randn_like(source_sub) → (48, 1000)
147
+ path_expr = AffineProbPath.sample(t_expr, noise_expr, target_sub)
148
+ path_expr.x_t = (1-t)*noise + t*target → (48, 1000) 插值后的噪声表达
149
+ path_expr.dx_t = target - noise → (48, 1000) 目标速度 (ground truth)
150
+
151
+ 潜变量流:
152
+ noise_latent = randn_like(z_target) → (48, 1000, 512)
153
+ 展平:
154
+ z_target_flat = z_target.reshape(48, 512000) → (48, 512000)
155
+ noise_latent_flat = noise_latent.reshape(48, 512000) → (48, 512000)
156
+ path_latent_flat = AffineProbPath.sample(t_latent, noise_latent_flat, z_target_flat)
157
+ 还原:
158
+ path_latent.x_t = reshape → (48, 1000, 512) 插值后的噪声 latent
159
+ path_latent.dx_t = reshape → (48, 1000, 512) 目标速度
160
+ ```
161
+
162
+ ### 2.5 模型前向传播 (`CascadedFlowModel.forward`)
163
+
164
+ ```
165
+ 输入:
166
+ gene_input: (48, 1000) 基因 token ID
167
+ source_sub: (48, 1000) 源表达
168
+ path_expr.x_t: (48, 1000) 噪声表达
169
+ path_latent.x_t: (48, 1000, 512) 噪声 latent
170
+ t_expr: (48,)
171
+ t_latent: (48,)
172
+ perturbation_id: (48, 2)
173
+ ```
174
+
175
+ #### Step 5a: 表达流 Embedding
176
+
177
+ ```
178
+ gene_emb = GeneEncoder(gene_input) → (48, 1000, 128)
179
+ val_emb_1 = ContinuousValueEncoder(x_t) → (48, 1000, 128) encoder_1 = 噪声 target (同 scDFM)
180
+ val_emb_2 = ContinuousValueEncoder(source_sub) → (48, 1000, 128) encoder_2 = control (同 scDFM)
181
+ expr_tokens = fusion_layer(cat[val_emb_1, val_emb_2]) + gene_emb
182
+ = Linear(256→128) → GELU → Linear(128→128) → LN + gene_emb → (48, 1000, 128)
183
+ ```
184
+
185
+ > **设计说明**(与 scDFM 对齐):
186
+ > - `value_encoder_1` 编码噪声 target(`x_t`),`value_encoder_2` 编码 control(`source`),与 scDFM 角色一致
187
+ > - `gene_emb` 只在融合后加一次(去冗余),而非分别加到两个 encoder 的输出上
188
+ > - `val_emb_2`(control 嵌入)传入 backbone 的 DiffPerceiverBlock 作为交叉注意力 KV(稳定参考基线)
189
+
190
+ #### Step 5b: 潜变量流 Embedding
191
+
192
+ ```
193
+ latent_tokens = LatentEmbedder(path_latent.x_t)
194
+ = Linear(512→128) → GELU → Linear(128→128) → (48, 1000, 128)
195
+ ```
196
+
197
+ #### Step 5c: 双流融合
198
+
199
+ ```
200
+ x = expr_tokens + latent_tokens → (48, 1000, 128)
201
+ ```
202
+
203
+ #### Step 5d: 条件向量
204
+
205
+ ```
206
+ t_expr_emb = TimestepEmbedder(t_expr) → (48, 128)
207
+ t_latent_emb = TimestepEmbedder(t_latent) → (48, 128)
208
+ pert_emb = GeneEncoder(perturbation_id).mean(dim=1) → (48, 128)
209
+ # perturbation_id: (48,2) → encoder → (48,2,128) → mean → (48,128)
210
+ c = t_expr_emb + t_latent_emb + pert_emb → (48, 128)
211
+ ```
212
+
213
+ #### Step 5e: 共享 Backbone (4 层 DiffPerceiverBlock)
214
+
215
+ ```
216
+ for i in range(4):
217
+ # GeneadaLN: 用 gene_emb 调制 x
218
+ x = gene_adaLN[i](gene_emb, x) → (48, 1000, 128)
219
+
220
+ # Adapter: 拼接 pert_emb 后降维
221
+ pert_exp = pert_emb[:, None, :].expand(-1, 1000, -1) → (48, 1000, 128)
222
+ x = cat[x, pert_exp] → (48, 1000, 256)
223
+ x = adapter_layer[i](x)
224
+ = Linear(256→128) → LeakyReLU → Linear(128→128) → LeakyReLU
225
+ → (48, 1000, 128)
226
+
227
+ # DiffPerceiverBlock: attention + MLP, 用 c 做 AdaLN 条件
228
+ x = DiffPerceiverBlock(x, val_emb_2, c) → (48, 1000, 128)
229
+ ```
230
+
231
+ #### Step 5f: 表达 Decoder Head
232
+
233
+ ```
234
+ x_with_pert = cat[x, pert_exp] → (48, 1000, 256)
235
+ pred_v_expr = ExprDecoder(x_with_pert)["pred"] → (48, 1000)
236
+ ```
237
+
238
+ #### Step 5g: Latent Decoder Head
239
+
240
+ ```
241
+ h = dh_proj(x) # Linear(128→128) → (48, 1000, 128)
242
+
243
+ for j in range(2): # dh_depth=2
244
+ # AdaLN: c → SiLU → Linear(128→768) → chunk 6
245
+ shift_msa, scale_msa, gate_msa,
246
+ shift_mlp, scale_mlp, gate_mlp = adaLN_modulation(c) 各 (48, 128)
247
+
248
+ # Self-Attention with AdaLN
249
+ h_norm = LayerNorm(h) * (1+scale_msa[:,None,:]) + shift_msa[:,None,:]
250
+ → (48, 1000, 128)
251
+ h_attn = MultiheadAttention(h_norm, h_norm, h_norm) → (48, 1000, 128)
252
+ h = h + gate_msa[:,None,:] * h_attn → (48, 1000, 128)
253
+
254
+ # MLP with AdaLN
255
+ h_norm = LayerNorm(h) * (1+scale_mlp[:,None,:]) + shift_mlp[:,None,:]
256
+ → (48, 1000, 128)
257
+ h_mlp = Linear(128→512) → GELU → Linear(512→128) → (48, 1000, 128)
258
+ h = h + gate_mlp[:,None,:] * h_mlp → (48, 1000, 128)
259
+
260
+ pred_v_latent = final(h)
261
+ = LN → Linear(128→128) → GELU → Linear(128→512) → (48, 1000, 512)
262
+ ```
263
+
264
+ ### 2.6 Loss 计算
265
+
266
+ ```
267
+ loss_expr = MSE(pred_v_expr - path_expr.dx_t) * w_expr[:, None]
268
+ = ((48,1000) - (48,1000))^2 * (48,1) → mean → scalar
269
+ 只对 w_expr=1 的样本有贡献 (约60%)
270
+
271
+ loss_latent = MSE(pred_v_latent - path_latent.dx_t) * w_latent[:, None, None]
272
+ = ((48,1000,512) - (48,1000,512))^2 * (48,1,1) → mean → scalar
273
+ 只对 w_latent=1 的样本有贡献 (约40%)
274
+
275
+ loss = loss_expr + latent_weight * loss_latent
276
+
277
+ 可选 MMD loss (对 w_expr>0 的样本):
278
+ x1_hat = x_t + pred_v_expr * (1-t_expr) # 单步重建 → (N_expr, 1000)
279
+ mmd_loss = mmd2_unbiased_multi_sigma(x1_hat, target_sub) → scalar
280
+ loss += gamma * mmd_loss
281
+ ```
282
+
283
+ ### 2.7 训练循环
284
+
285
+ ```
286
+ for iteration in range(200000):
287
+ batch → source(48, 5000), target(48, 5000), pert_id(48, 2)
288
+ loss = denoiser.train_step(source, target, pert_id, gene_ids, infer_top_gene=1000)
289
+ optimizer.zero_grad()
290
+ accelerator.backward(loss)
291
+ optimizer.step()
292
+ scheduler.step() # CosineAnnealing
293
+
294
+ if iteration % 5000 == 0:
295
+ save checkpoint + run evaluation
296
+ ```
297
+
298
+ ---
299
+
300
+ ## 三、推理流程 Tensor 数据流
301
+
302
+ 推理时**不调用 scGPT 编码器**,模型从噪声自行生成 latent(辅助语义表征)。用**全部 5000 个基因**,不做子集采样。
303
+ 条件信号(control 表达 + perturbation_id)在每个 ODE 步中持续提供。
304
+
305
+ ```
306
+ 输入:
307
+ source: (B, 5000) 控制组表达
308
+ perturbation_id: (B, 2) 扰动 ID
309
+ gene_ids: (B, 5000) 基因 vocab ID
310
+ ```
311
+
312
+ ### 3.1 Stage 1: 生成 Latent(t_latent: 0→1, t_expr 固定=0)
313
+
314
+ ```
315
+ z_noise = randn(B, 5000, 512) 初始噪声
316
+
317
+ ODE 求解 (RK4, 20步, t ∈ [0,1]):
318
+ 每一步调用 latent_vf(t, z_t):
319
+ t_latent = t.expand(B) → (B,) 当前时间步
320
+ t_expr = zeros(B) → (B,) 固定为 0
321
+
322
+ model.forward(
323
+ gene_ids, (B, 5000)
324
+ source, (B, 5000) 源表达
325
+ source, (B, 5000) ← 注意: x_t 用 source (因为 t_expr=0)
326
+ z_t, (B, 5000, 512) 当前 latent 状态
327
+ t_expr=0, (B,)
328
+ t_latent=t, (B,)
329
+ perturbation_id, (B, 2)
330
+ )
331
+ → _, v_latent (B, 5000, 512) latent 速度场
332
+
333
+ return v_latent
334
+
335
+ z_traj = odeint(latent_vf, z_noise, linspace(0,1,20))
336
+ → (20, B, 5000, 512)
337
+ z_generated = z_traj[-1] → (B, 5000, 512) ★ 生成的 latent
338
+ ```
339
+
340
+ ### 3.2 Stage 2: 用生成的 Latent 引导表达生成(t_expr: 0→1, t_latent 固定=1)
341
+
342
+ ```
343
+ expr_noise = randn_like(source) → (B, 5000) 高斯噪声 (或 Poisson)
344
+
345
+ ODE 求解 (RK4, 20步, t ∈ [0,1]):
346
+ 每一步调用 expr_vf(t, x_t):
347
+ t_expr = t.expand(B) → (B,) 当前时间步
348
+ t_latent = ones(B) → (B,) 固定为 1 (latent 已完成)
349
+
350
+ model.forward(
351
+ gene_ids, (B, 5000)
352
+ source, (B, 5000)
353
+ x_t, (B, 5000) ← 当前噪声表达
354
+ z_generated, (B, 5000, 512) ★ Stage 1 的结果 (固定)
355
+ t_expr=t, (B,)
356
+ t_latent=1, (B,)
357
+ perturbation_id, (B, 2)
358
+ )
359
+ → v_expr, _ (B, 5000) 表达速度场
360
+
361
+ return v_expr
362
+
363
+ x_traj = odeint(expr_vf, expr_noise, linspace(0,1,20))
364
+ → (20, B, 5000)
365
+ x_generated = clamp(x_traj[-1], min=0) → (B, 5000) ★ 最终预测表达
366
+ ```
367
+
368
+ ### 3.3 评估流程
369
+
370
+ ```
371
+ 对每个扰动条件:
372
+ source = 随机采样 128 个控制细胞 → (128, 5000)
373
+ 按 batch_size 调用 generate:
374
+ pred = denoiser.generate(source, pert_id, gene_ids) → (128, 5000)
375
+
376
+ 汇总:
377
+ pred_adata = AnnData(X=所有预测表达, obs=扰动标签)
378
+ real_adata = AnnData(X=所有真实表达, obs=扰动标签)
379
+ MetricsEvaluator(pred_adata, real_adata) → results.csv, agg_results.csv
380
+ ```
381
+
382
+ ---
383
+
384
+ ## ��、训练 vs 推理 对比总结
385
+
386
+ | 维度 | 训练 | 推理 |
387
+ |---|---|---|
388
+ | 基因数 | 随机采样 G=1000 | 全量 G=5000 |
389
+ | scGPT 特征(辅助生成目标) | 冻结提取 target 特征 `z_target` 作为 latent 流的生成目标 | **不使用**,模型从噪声自行生成 latent(Stage 1) |
390
+ | 时间步 | 级联采样:40% 训练 latent, 60% 训练表达 | 两阶段串行:先 latent(0→1), 再表达(0→1) |
391
+ | 流采样 | 单步:`x_t = (1-t)*noise + t*target` | ODE 积分:RK4, 20 步 |
392
+ | Loss | MSE(velocity) + MMD | 无 |
393
+ | 输出 | scalar loss | `(B, G_full)` 表达矩阵 |
394
+ | 条件信号 | control + pert_id(真条件);scGPT 特征是生成目标 | control + pert_id(每步输入);latent 由 Stage1 生成 |
395
+ | 时间步 | `t_expr` 和 `t_latent` 随机,互斥激活 | Stage1: `t_expr=0, t_latent∈[0,1]`; Stage2: `t_latent=1, t_expr∈[0,1]` |
396
+
397
+ ---
398
+
399
+ ## 五、模型架构总览图
400
+
401
+ ```
402
+ ┌──────────────────────────────────────────────┐
403
+ │ CascadedFlowModel │
404
+ │ │
405
+ gene_id (B,G) ───────→│ GeneEncoder ──→ gene_emb (B,G,d) │
406
+ │ │ │
407
+ x_t (B,G) ──────────→│ ValEnc1 ──→ val_emb_1 (B,G,d) 噪声target │
408
+ source (B,G) ────────→│ ValEnc2 ──→ val_emb_2 (B,G,d) control │
409
+ │ │ │
410
+ │ fusion_layer(cat[val_emb_1, val_emb_2]) │
411
+ │ + gene_emb (融合后加一次) │
412
+ │ ↓ │
413
+ │ expr_tokens (B,G,d) │
414
+ │ │ │
415
+ z_t (B,G,512) ──────→│ LatentEmbedder(512→128→d) │
416
+ │ ↓ │
417
+ │ latent_tokens (B,G,d) │
418
+ │ │ │
419
+ │ x = expr_tokens + latent_tokens (B,G,d) │
420
+ │ │
421
+ t_expr (B,) ────────→│ TimestepEmb ──┐ │
422
+ t_latent (B,) ──────→│ TimestepEmb ──┼→ c = sum (B,d) │
423
+ pert_id (B,2) ──────→│ PertEmb ──────┘ │
424
+ │ │
425
+ │ ┌─ 4x DiffPerceiverBlock ──────────────┐ │
426
+ │ │ gene_adaLN(gene_emb, x) │ │
427
+ │ │ cat[x, pert_emb] → adapter → x │ │
428
+ │ │ DiffPerceiverBlock(x, val_emb_2, c) │ │
429
+ │ └──────────────────────────────────────┘ │
430
+ │ │ │
431
+ │ ├──→ ExprDecoder(cat[x, pert]) ──→ pred_v_expr (B,G)
432
+ │ │ │
433
+ │ └──→ LatentDecoder(x, c) │
434
+ │ dh_proj → 2x AdaLN Block │
435
+ │ → final proj ──→ pred_v_latent (B,G,512)
436
+ └──────────────────────────────────────────────┘
437
+ ```
transfer/code/CCFM/_bootstrap_scdfm.py ADDED
@@ -0,0 +1,101 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Bootstrap scDFM imports by creating missing __init__.py files and loading
3
+ its modules under a 'scdfm_src' prefix in sys.modules.
4
+
5
+ This module MUST be imported before any CCFM src imports.
6
+ """
7
+
8
+ import sys
9
+ import os
10
+ import types
11
+
12
+ _SCDFM_ROOT = os.path.normpath(
13
+ os.path.join(os.path.dirname(os.path.abspath(__file__)), "..", "scDFM")
14
+ )
15
+
16
+ # Directories in scDFM that need __init__.py to be proper packages
17
+ _DIRS_NEEDING_INIT = [
18
+ "src",
19
+ "src/models",
20
+ "src/models/origin",
21
+ "src/data_process",
22
+ "src/tokenizer",
23
+ "src/script",
24
+ "src/models/perturbation",
25
+ ]
26
+
27
+
28
+ def _ensure_init_files():
29
+ """Create missing __init__.py files in scDFM so it can be imported as packages."""
30
+ created = []
31
+ for d in _DIRS_NEEDING_INIT:
32
+ init_path = os.path.join(_SCDFM_ROOT, d, "__init__.py")
33
+ if not os.path.exists(init_path):
34
+ with open(init_path, "w") as f:
35
+ f.write("# Auto-created by CCFM bootstrap\n")
36
+ created.append(init_path)
37
+ return created
38
+
39
+
40
+ def bootstrap():
41
+ """Load scDFM's src package as 'scdfm_src' in sys.modules."""
42
+ if "scdfm_src" in sys.modules:
43
+ return # Already bootstrapped
44
+
45
+ # Create missing __init__.py files
46
+ _ensure_init_files()
47
+
48
+ # Save CCFM's src modules
49
+ saved = {}
50
+ for key in list(sys.modules.keys()):
51
+ if key == "src" or key.startswith("src."):
52
+ saved[key] = sys.modules.pop(key)
53
+
54
+ # Add scDFM root to path
55
+ sys.path.insert(0, _SCDFM_ROOT)
56
+
57
+ try:
58
+ # Import scDFM modules (their relative imports work now)
59
+ import src as scdfm_src_pkg
60
+ import src.models
61
+ import src.models.origin
62
+ import src.models.origin.blocks
63
+ import src.models.origin.layers
64
+ import src.models.origin.model
65
+ import src.flow_matching
66
+ import src.flow_matching.path
67
+ import src.flow_matching.path.path
68
+ import src.flow_matching.path.path_sample
69
+ import src.flow_matching.path.affine
70
+ import src.flow_matching.path.scheduler
71
+ import src.flow_matching.path.scheduler.scheduler
72
+ # Skip src.flow_matching.ot (requires 'ot' package, not needed for CCFM)
73
+ import src.utils
74
+ import src.utils.utils
75
+ import src.tokenizer
76
+ import src.tokenizer.gene_tokenizer
77
+ # Skip src.data_process (has heavy deps like bs4, rdkit)
78
+ # We handle data loading separately in CCFM
79
+
80
+ # Re-register all under scdfm_src.* prefix
81
+ for key in list(sys.modules.keys()):
82
+ if key == "src" or key.startswith("src."):
83
+ new_key = "scdfm_" + key
84
+ sys.modules[new_key] = sys.modules[key]
85
+
86
+ finally:
87
+ # Remove scDFM's src.* entries
88
+ for key in list(sys.modules.keys()):
89
+ if (key == "src" or key.startswith("src.")) and not key.startswith("scdfm_"):
90
+ del sys.modules[key]
91
+
92
+ # Restore CCFM's src modules
93
+ for key, mod in saved.items():
94
+ sys.modules[key] = mod
95
+
96
+ # Remove scDFM from front of path
97
+ if _SCDFM_ROOT in sys.path:
98
+ sys.path.remove(_SCDFM_ROOT)
99
+
100
+
101
+ bootstrap()
transfer/code/CCFM/cell-eval-outdir/agg_results.csv ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ statistic,overlap_at_N,overlap_at_50,overlap_at_100,overlap_at_200,overlap_at_500,precision_at_N,precision_at_50,precision_at_100,precision_at_200,precision_at_500,de_spearman_sig,de_direction_match,de_spearman_lfc_sig,de_sig_genes_recall,de_nsig_counts_real,de_nsig_counts_pred,pr_auc,roc_auc,pearson_delta,mse,mae,mse_delta,mae_delta,discrimination_score_l1,discrimination_score_l2,discrimination_score_cosine,pearson_edistance,clustering_agreement
2
+ count,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0
3
+ null_count,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
4
+ mean,0.07395320375987084,0.06849358974358978,0.0725404899006272,0.07431250785187005,0.07395320375987084,0.12877704205127136,0.06871794871794874,0.07282051282051284,0.07500000000000001,0.06682051282051284,0.0006180401974148523,0.7319719710281621,0.7355229056966586,0.9061342423642376,137.89743589743588,977.6153846153846,0.21954205885448827,0.502651350313865,-0.08164090049789667,0.03866601647089939,0.054930761086430545,0.03866601647089939,0.054930761086430545,0.5141354372123603,0.5121630506245891,0.5115055884286653,0.12324841081988779,0.007165149729861788
5
+ std,0.06327073571923675,0.04142623783921817,0.06430260938703716,0.06497823680683154,0.06327073571923675,0.0713575372552735,0.04124251046390996,0.06398802182105803,0.06369127842466676,0.06055657285150447,1.0983753514559728e-19,0.05548309801759892,0.06748831752084138,0.027067878739138167,72.75871981447884,2.2197129332580094,0.07449489082434584,0.025106101263114602,0.11079253285523155,0.01804318670820533,0.013893820583993645,0.01804318670820533,0.013893820583993646,0.2932360142364609,0.29232232244842243,0.30752068526485193,2.8118408997272905e-17,2.636100843494335e-18
6
+ min,0.0,0.0,0.0,0.0,0.0,0.030706243602865915,0.0,0.0,0.005,0.008,0.0006180401974148526,0.6301369863013698,0.5466766243465273,0.8247422680412371,32.0,972.0,0.07528683946739989,0.45502409257502185,-0.25674855274153263,0.017089386967776357,0.041686743621341026,0.017089386967776357,0.041686743621341026,0.02564102564102566,0.02564102564102566,0.02564102564102566,0.12324841081988787,0.007165149729861791
7
+ 25%,0.042105263157894736,0.04,0.04,0.042105263157894736,0.042105263157894736,0.09397344228804903,0.04,0.04,0.045,0.038,0.0006180401974148526,0.7037037037037037,0.6922532601004144,0.8904109589041096,103.0,976.0,0.17885902518604974,0.48868455014749274,-0.16340383087297497,0.029862651934889305,0.048704704053449004,0.029862651934889305,0.048704704053449004,0.3076923076923077,0.3076923076923077,0.2564102564102564,0.12324841081988787,0.007165149729861791
8
+ 50%,0.05454545454545454,0.06,0.0547945205479452,0.05454545454545454,0.05454545454545454,0.10827374872318693,0.06,0.05,0.055,0.048,0.0006180401974148526,0.7272727272727273,0.7283894176721877,0.9029126213592233,117.0,978.0,0.2155530179012921,0.5015608589100246,-0.10351595518920989,0.03463771528656705,0.05123648361160402,0.03463771528656705,0.05123648361160402,0.5128205128205128,0.5384615384615384,0.5128205128205128,0.12324841081988787,0.007165149729861791
9
+ 75%,0.08333333333333333,0.1,0.08,0.08333333333333333,0.08333333333333333,0.14974358974358976,0.1,0.08,0.08,0.068,0.0006180401974148526,0.7551020408163265,0.793306240068219,0.9259259259259259,158.0,979.0,0.2577585712884973,0.5154239019407559,0.007450254288251119,0.04417537517338548,0.0574020448039485,0.04417537517338548,0.0574020448039485,0.7435897435897436,0.7692307692307692,0.7948717948717949,0.12324841081988787,0.007165149729861791
10
+ max,0.30548302872062666,0.16,0.34,0.33,0.30548302872062666,0.3719262295081967,0.16,0.34,0.33,0.3,0.0006180401974148526,0.8705882352941177,0.8543173721436267,0.9506172839506173,384.0,983.0,0.41857905891786573,0.5534962262791009,0.22806546181808088,0.11179507217860543,0.1123107911192389,0.11179507217860543,0.1123107911192389,1.0,1.0,1.0,0.12324841081988787,0.007165149729861791
transfer/code/CCFM/cell-eval-outdir/pred_de.csv ADDED
The diff for this file is too large to render. See raw diff
 
transfer/code/CCFM/cell-eval-outdir/real_de.csv ADDED
The diff for this file is too large to render. See raw diff
 
transfer/code/CCFM/cell-eval-outdir/results.csv ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ perturbation,overlap_at_N,overlap_at_50,overlap_at_100,overlap_at_200,overlap_at_500,precision_at_N,precision_at_50,precision_at_100,precision_at_200,precision_at_500,de_spearman_sig,de_direction_match,de_spearman_lfc_sig,de_sig_genes_recall,de_nsig_counts_real,de_nsig_counts_pred,pr_auc,roc_auc,pearson_delta,mse,mae,mse_delta,mae_delta,discrimination_score_l1,discrimination_score_l2,discrimination_score_cosine,pearson_edistance,clustering_agreement
2
+ AHR+FEV,0.1308411214953271,0.14,0.14,0.125,0.1308411214953271,0.20245398773006135,0.14,0.14,0.125,0.1,0.0006180401974148526,0.7897196261682243,0.793306240068219,0.9252336448598131,214.0,978.0,0.289355620021763,0.5044885971796151,-0.000681255195205219,0.05580747882233428,0.0703616503153296,0.05580747882233427,0.0703616503153296,0.10256410256410253,0.15384615384615385,0.7948717948717949,0.12324841081988787,0.007165149729861791
3
+ AHR+KLF1,0.1,0.1,0.09,0.1,0.1,0.12589559877175024,0.1,0.09,0.085,0.068,0.0006180401974148526,0.6714285714285714,0.7230217243868823,0.8785714285714286,140.0,977.0,0.2155530179012921,0.5011627906976744,-0.17882777360167185,0.03415252740506693,0.053647045372835565,0.03415252740506693,0.053647045372835565,0.3076923076923077,0.46153846153846156,0.10256410256410253,0.12324841081988787,0.007165149729861791
4
+ BCL2L11+BAK1,0.03125,0.03125,0.03125,0.03125,0.03125,0.030706243602865915,0.04,0.05,0.045,0.03,0.0006180401974148526,0.75,0.8320806599450046,0.9375,32.0,977.0,0.07528683946739989,0.5477143595041322,0.007450254288251119,0.018631415934436527,0.042142537012775594,0.018631415934436527,0.042142537012775594,0.9487179487179487,0.9487179487179487,0.7948717948717949,0.12324841081988787,0.007165149729861791
5
+ BPGM+ZBTB1,0.05454545454545454,0.08,0.06,0.05454545454545454,0.05454545454545454,0.10601427115188583,0.08,0.06,0.05,0.04,0.0006180401974148526,0.7272727272727273,0.7216076912323739,0.9454545454545454,110.0,981.0,0.15723145661369453,0.4712257405515833,-0.14295966613751018,0.03777198411559816,0.05109172018229229,0.03777198411559816,0.05109172018229229,0.5128205128205128,0.3846153846153846,0.33333333333333337,0.12324841081988787,0.007165149729861791
6
+ CBL+PTPN12,0.04807692307692308,0.06,0.05,0.04807692307692308,0.04807692307692308,0.09397344228804903,0.06,0.05,0.05,0.03,0.0006180401974148526,0.7115384615384616,0.5466766243465273,0.8846153846153846,104.0,979.0,0.2527906962637153,0.5022428743131868,-0.18550991579880255,0.04731173214331579,0.05330420638914829,0.04731173214331579,0.05330420638914829,0.3846153846153846,0.23076923076923073,0.17948717948717952,0.12324841081988787,0.007165149729861791
7
+ CBL+PTPN9,0.042105263157894736,0.02,0.042105263157894736,0.042105263157894736,0.042105263157894736,0.08709016393442623,0.02,0.04,0.05,0.032,0.0006180401974148526,0.7157894736842105,0.6192082645090848,0.8947368421052632,95.0,976.0,0.23347166134256753,0.5006746147135795,-0.11418825671976725,0.0345185799724352,0.048704704053449004,0.0345185799724352,0.048704704053449004,0.717948717948718,0.5897435897435898,0.46153846153846156,0.12324841081988787,0.007165149729861791
8
+ CBL+TGFBR2,0.0547945205479452,0.04,0.0547945205479452,0.0547945205479452,0.0547945205479452,0.06659836065573771,0.04,0.04,0.05,0.038,0.0006180401974148526,0.6301369863013698,0.6730638691761802,0.8904109589041096,73.0,976.0,0.1657458038482859,0.5353327126834242,-0.10350698889045816,0.03266801215509963,0.04836069691089276,0.03266801215509963,0.04836069691089276,0.7435897435897436,0.641025641025641,0.5897435897435898,0.12324841081988787,0.007165149729861791
9
+ CBL+UBASH3A,0.0,0.0,0.0,0.0,0.0,0.043209876543209874,0.0,0.0,0.005,0.008,0.0006180401974148526,0.6595744680851063,0.8201896392229417,0.8936170212765957,47.0,972.0,0.12664611232108428,0.5477662923355139,-0.16340383087297497,0.03660779070416713,0.05174531782519213,0.03660779070416713,0.05174531782519213,0.5128205128205128,0.5384615384615384,0.2564102564102564,0.12324841081988787,0.007165149729861791
10
+ CBL+UBASH3B,0.08904109589041095,0.14,0.1,0.08904109589041095,0.08904109589041095,0.13190184049079753,0.14,0.1,0.07,0.058,0.0006180401974148526,0.7328767123287672,0.6719708226394705,0.8835616438356164,146.0,978.0,0.2927739461231338,0.5109236148984634,-0.09303837127636658,0.047811851528487485,0.05215327697445609,0.047811851528487485,0.05215327697445609,0.3589743589743589,0.20512820512820518,0.4871794871794872,0.12324841081988787,0.007165149729861791
11
+ CDKN1B+CDKN1A,0.02912621359223301,0.02,0.03,0.02912621359223301,0.02912621359223301,0.09489795918367347,0.02,0.03,0.06,0.048,0.0006180401974148526,0.7572815533980582,0.7994003426613364,0.9029126213592233,103.0,980.0,0.13727131313018917,0.4975701096427142,-0.15707749719183386,0.029985238722000304,0.05206961946948075,0.029985238722000304,0.05206961946948075,0.46153846153846156,0.641025641025641,0.33333333333333337,0.12324841081988787,0.007165149729861791
12
+ CDKN1C+CDKN1B,0.03636363636363636,0.04,0.03,0.03636363636363636,0.03636363636363636,0.1023541453428864,0.04,0.03,0.045,0.042,0.0006180401974148526,0.7090909090909091,0.8149440522239354,0.9090909090909091,110.0,977.0,0.15180512654084866,0.5154239019407559,-0.14115175381435643,0.029967122965036647,0.05138300370287098,0.029967122965036647,0.05138300370287098,0.5897435897435898,0.6923076923076923,0.5128205128205128,0.12324841081988787,0.007165149729861791
13
+ CEBPE+CEBPA,0.30548302872062666,0.1,0.34,0.33,0.30548302872062666,0.3707865168539326,0.1,0.34,0.33,0.3,0.0006180401974148526,0.8694516971279374,0.729553310592033,0.9477806788511749,383.0,979.0,0.4157540694520757,0.49803648581741855,0.04068628601729422,0.11179507217860543,0.1123107911192389,0.11179507217860543,0.1123107911192389,0.02564102564102566,0.02564102564102566,0.9230769230769231,0.12324841081988787,0.007165149729861791
14
+ CEBPE+CNN1,0.11564625850340136,0.14,0.12,0.11564625850340136,0.11564625850340136,0.14096016343207354,0.14,0.12,0.125,0.082,0.0006180401974148526,0.7551020408163265,0.8324352511733009,0.9387755102040817,147.0,979.0,0.23976185026595853,0.5499198507069885,-0.04180838169993791,0.02799969093443837,0.05105374262657179,0.02799969093443837,0.05105374262657179,0.5897435897435898,0.7948717948717949,0.6923076923076923,0.12324841081988787,0.007165149729861791
15
+ ETS2+IGDCC3,0.06956521739130435,0.06,0.06,0.06956521739130435,0.06956521739130435,0.10633946830265849,0.06,0.06,0.075,0.058,0.0006180401974148526,0.7391304347826086,0.7768612654710786,0.9043478260869565,115.0,978.0,0.23370533925403658,0.495897813804962,0.029532205030916036,0.020717920454122067,0.045982380799085745,0.020717920454122067,0.045982380799085745,0.8205128205128205,0.8205128205128205,0.8205128205128205,0.12324841081988787,0.007165149729861791
16
+ ETS2+IKZF3,0.10752688172043011,0.12,0.11,0.10752688172043011,0.10752688172043011,0.17008196721311475,0.12,0.11,0.1,0.086,0.0006180401974148526,0.7634408602150538,0.7704231330786419,0.8924731182795699,186.0,976.0,0.2453396755066589,0.48778764101344746,-0.18685575986849334,0.05693390716612494,0.060505517368878346,0.05693390716612494,0.060505517368878346,0.1282051282051282,0.1282051282051282,0.07692307692307687,0.12324841081988787,0.007165149729861791
17
+ FEV+ISL2,0.05263157894736842,0.04,0.04,0.05263157894736842,0.05263157894736842,0.13655030800821355,0.04,0.04,0.065,0.066,0.0006180401974148526,0.7105263157894737,0.7411408742029778,0.875,152.0,974.0,0.1802385047469924,0.46315246400198606,-0.08184300427470183,0.04140964303946295,0.05993504252580628,0.04140964303946295,0.05993504252580628,0.20512820512820518,0.2564102564102564,0.41025641025641024,0.12324841081988787,0.007165149729861791
18
+ FOSB+CEBPB,0.1803921568627451,0.16,0.19,0.185,0.1803921568627451,0.24487704918032788,0.16,0.19,0.185,0.15,0.0006180401974148526,0.8705882352941177,0.7788869424215662,0.9372549019607843,255.0,976.0,0.2895674815827421,0.49720226345571794,0.07216521168170856,0.06788741332834797,0.08781863929818466,0.06788741332834797,0.08781863929818468,0.07692307692307687,0.05128205128205132,0.8461538461538461,0.12324841081988787,0.007165149729861791
19
+ FOXA3+FOXA1,0.07857142857142857,0.1,0.08,0.07857142857142857,0.07857142857142857,0.12883435582822086,0.1,0.08,0.075,0.064,0.0006180401974148526,0.6857142857142857,0.8037918215613383,0.9,140.0,978.0,0.17885902518604974,0.4941901993355481,-0.05074626553588181,0.029862651934889305,0.05398229729784066,0.029862651934889305,0.05398229729784066,0.33333333333333337,0.7692307692307692,0.641025641025641,0.12324841081988787,0.007165149729861791
20
+ KLF1+BAK1,0.052083333333333336,0.1,0.052083333333333336,0.052083333333333336,0.052083333333333336,0.08597748208802457,0.1,0.05,0.05,0.042,0.0006180401974148526,0.7083333333333334,0.6907398360744721,0.875,96.0,977.0,0.19735133996261323,0.48868455014749274,-0.24515315071378704,0.03197165400949463,0.04877062774155349,0.03197165400949463,0.04877062774155349,0.7435897435897436,0.6666666666666667,0.02564102564102566,0.12324841081988787,0.007165149729861791
21
+ KLF1+CEBPA,0.2942708333333333,0.1,0.25,0.285,0.2942708333333333,0.3719262295081967,0.1,0.25,0.285,0.298,0.0006180401974148526,0.8671875,0.7524491829127294,0.9453125,384.0,976.0,0.41857905891786573,0.5074277935606061,0.033241542515528716,0.06595129422363945,0.08905813105074563,0.06595129422363945,0.08905813105074563,0.05128205128205132,0.07692307692307687,0.8717948717948718,0.12324841081988787,0.007165149729861791
22
+ KLF1+CLDN6,0.05343511450381679,0.04,0.04,0.05343511450381679,0.05343511450381679,0.12040816326530612,0.04,0.04,0.055,0.056,0.0006180401974148526,0.7022900763358778,0.7190155882980995,0.9007633587786259,131.0,980.0,0.21494914880766414,0.5019984363882325,-0.25674855274153263,0.051408548351699966,0.059620483832245184,0.051408548351699966,0.059620483832245184,0.17948717948717952,0.17948717948717952,0.05128205128205132,0.12324841081988787,0.007165149729861791
23
+ LYL1+CEBPB,0.0949367088607595,0.06,0.08,0.0949367088607595,0.0949367088607595,0.14974358974358976,0.06,0.08,0.095,0.078,0.0006180401974148526,0.7848101265822784,0.8251475211277035,0.9240506329113924,158.0,975.0,0.218219692606571,0.504111669021919,-0.16514919028268413,0.029257650899615014,0.053947390749385,0.029257650899615014,0.053947390749385,0.3076923076923077,0.6666666666666667,0.1282051282051282,0.12324841081988787,0.007165149729861791
24
+ MAP2K3+ELMSAN1,0.025974025974025976,0.0,0.02,0.025974025974025976,0.025974025974025976,0.14504596527068436,0.0,0.02,0.04,0.06,0.0006180401974148526,0.7272727272727273,0.718742031589043,0.922077922077922,154.0,979.0,0.2577585712884973,0.48963418378311996,0.020462538397928824,0.019811996728230724,0.04367858584334204,0.019811996728230724,0.04367858584334204,0.8974358974358975,0.8461538461538461,0.8205128205128205,0.12324841081988787,0.007165149729861791
25
+ MAP2K3+IKZF3,0.08333333333333333,0.1,0.08,0.08333333333333333,0.08333333333333333,0.09805924412665985,0.1,0.08,0.055,0.046,0.0006180401974148526,0.7037037037037037,0.7167014593878388,0.8888888888888888,108.0,979.0,0.22899973834921372,0.4874553645573824,-0.20767238076053746,0.0652399118899808,0.06086253874483393,0.0652399118899808,0.06086253874483393,0.20512820512820518,0.10256410256410253,0.1282051282051282,0.12324841081988787,0.007165149729861791
26
+ MAP2K3+MAP2K6,0.028169014084507043,0.04,0.028169014084507043,0.028169014084507043,0.028169014084507043,0.065439672801636,0.04,0.03,0.025,0.034,0.0006180401974148526,0.6901408450704225,0.6922532601004144,0.9014084507042254,71.0,978.0,0.130106127841948,0.5246061947573492,0.10542488939536575,0.017089386967776357,0.042199752748710434,0.017089386967776357,0.042199752748710434,0.9743589743589743,0.9743589743589743,0.9743589743589743,0.12324841081988787,0.007165149729861791
27
+ MAP2K3+SLC38A2,0.05263157894736842,0.04,0.05263157894736842,0.05263157894736842,0.05263157894736842,0.051934826883910386,0.04,0.05,0.035,0.032,0.0006180401974148526,0.7017543859649122,0.6621428571428571,0.8947368421052632,57.0,982.0,0.09447034806783824,0.45502409257502185,0.22806546181808088,0.01764320931428523,0.041686743621341026,0.01764320931428523,0.041686743621341026,1.0,1.0,1.0,0.12324841081988787,0.007165149729861791
28
+ MAP2K6+ELMSAN1,0.046511627906976744,0.02,0.046511627906976744,0.046511627906976744,0.046511627906976744,0.07865168539325842,0.02,0.05,0.055,0.046,0.0006180401974148526,0.7558139534883721,0.7393594426981565,0.8953488372093024,86.0,979.0,0.14357163096260944,0.46603226298916084,0.07308970049844386,0.01884736438588656,0.04362012745480605,0.01884736438588656,0.04362012745480605,0.9487179487179487,0.9230769230769231,0.9487179487179487,0.12324841081988787,0.007165149729861791
29
+ MAPK1+IKZF3,0.07317073170731707,0.04,0.04,0.07317073170731707,0.07317073170731707,0.15462868769074262,0.04,0.04,0.075,0.058,0.0006180401974148526,0.725609756097561,0.7737163223801006,0.926829268292683,164.0,983.0,0.27474306161131856,0.5015608589100246,-0.18217857153146724,0.0425078665820119,0.0574020448039485,0.0425078665820119,0.0574020448039485,0.2564102564102564,0.3076923076923077,0.1282051282051282,0.12324841081988787,0.007165149729861791
30
+ PTPN12+PTPN9,0.061224489795918366,0.06,0.061224489795918366,0.061224489795918366,0.061224489795918366,0.0869120654396728,0.06,0.06,0.045,0.03,0.0006180401974148526,0.7142857142857143,0.6621602663027536,0.8673469387755102,98.0,978.0,0.25011674665005473,0.5097402597402598,-0.20268265319102213,0.03914455888068411,0.05023655449671707,0.03914455888068411,0.05023655449671707,0.6923076923076923,0.4358974358974359,0.20512820512820518,0.12324841081988787,0.007165149729861791
31
+ PTPN12+UBASH3A,0.06086956521739131,0.1,0.07,0.06086956521739131,0.06086956521739131,0.10245901639344263,0.1,0.07,0.065,0.048,0.0006180401974148526,0.7304347826086957,0.6872912173681067,0.8695652173913043,115.0,976.0,0.20163301499093328,0.4667747482191108,-0.1405571263181629,0.04187065351581521,0.05191913093302079,0.041870653515815204,0.05191913093302079,0.41025641025641024,0.33333333333333337,0.2564102564102564,0.12324841081988787,0.007165149729861791
32
+ SGK1+S1PR2,0.08205128205128205,0.08,0.07,0.08205128205128205,0.08205128205128205,0.18692543411644535,0.08,0.07,0.085,0.104,0.0006180401974148526,0.7487179487179487,0.7726498750939433,0.9384615384615385,195.0,979.0,0.27246186996366134,0.504405160057334,-0.1425496176526846,0.04362755416454579,0.061769647398377175,0.04362755416454579,0.061769647398377175,0.15384615384615385,0.3589743589743589,0.3589743589743589,0.12324841081988787,0.007165149729861791
33
+ SGK1+TBX2,0.06557377049180328,0.1,0.06,0.06557377049180328,0.06557377049180328,0.11349693251533742,0.1,0.06,0.065,0.052,0.0006180401974148526,0.7295081967213115,0.7230355295385358,0.9098360655737705,122.0,978.0,0.23721339749548936,0.5243194294036371,-0.17427454250530847,0.04053582957303282,0.0501051497886934,0.04053582957303282,0.0501051497886934,0.5641025641025641,0.41025641025641024,0.20512820512820518,0.12324841081988787,0.007165149729861791
34
+ SGK1+TBX3,0.045871559633027525,0.08,0.05,0.045871559633027525,0.045871559633027525,0.10245901639344263,0.08,0.05,0.025,0.024,0.0006180401974148526,0.6880733944954128,0.6942678111269721,0.9174311926605505,109.0,976.0,0.3383204059729327,0.5534962262791009,-0.08464092901325047,0.04417537517338548,0.05123648361160402,0.04417537517338548,0.05123648361160402,0.5641025641025641,0.3076923076923077,0.6666666666666667,0.12324841081988787,0.007165149729861791
35
+ TBX3+TBX2,0.030864197530864196,0.02,0.03,0.030864197530864196,0.030864197530864196,0.15337423312883436,0.02,0.03,0.03,0.066,0.0006180401974148526,0.7530864197530864,0.7208131055895328,0.9259259259259259,162.0,978.0,0.2846346150206432,0.5430920180323522,-0.019667607914308216,0.03463771528656705,0.04748869970105704,0.03463771528656705,0.04748869970105704,0.7948717948717949,0.5384615384615384,0.7692307692307692,0.12324841081988787,0.007165149729861791
36
+ TGFBR2+IGDCC3,0.03669724770642202,0.06,0.04,0.03669724770642202,0.03669724770642202,0.09908069458631256,0.06,0.04,0.04,0.03,0.0006180401974148526,0.6330275229357798,0.6363139235813425,0.8899082568807339,109.0,979.0,0.18103811865440816,0.4730485280943997,-0.10351595518920989,0.031240376498607152,0.05090387243412622,0.031240376498607152,0.05090387243412622,0.6153846153846154,0.7692307692307692,0.6153846153846154,0.12324841081988787,0.007165149729861791
37
+ TGFBR2+PRTG,0.03418803418803419,0.04,0.04,0.03418803418803419,0.03418803418803419,0.10827374872318693,0.04,0.04,0.045,0.046,0.0006180401974148526,0.6752136752136753,0.7283894176721877,0.905982905982906,117.0,979.0,0.1710995969889653,0.4782598174444153,-0.14178338892395337,0.03324635981127433,0.05022525364973255,0.03324635981127433,0.05022525364973255,0.46153846153846156,0.5641025641025641,0.5128205128205128,0.12324841081988787,0.007165149729861791
38
+ UBASH3B+OSR2,0.04580152671755725,0.1,0.06,0.04580152671755725,0.04580152671755725,0.12397540983606557,0.1,0.06,0.05,0.046,0.0006180401974148526,0.732824427480916,0.8088719153936546,0.9236641221374046,131.0,976.0,0.18232471960433552,0.5158161965582972,-0.02485046822501524,0.02185241151506355,0.044643295457333705,0.02185241151506355,0.044643295457333705,0.8974358974358975,0.8717948717948718,0.717948717948718,0.12324841081988787,0.007165149729861791
39
+ UBASH3B+ZBTB25,0.010309278350515464,0.02,0.010309278350515464,0.010309278350515464,0.010309278350515464,0.08221993833504625,0.02,0.02,0.035,0.032,0.0006180401974148526,0.711340206185567,0.6584528577347248,0.8247422680412371,97.0,973.0,0.18033908245418523,0.47326209313742285,-0.1533095016375446,0.036293167827057715,0.04955475749838851,0.036293167827057715,0.04955475749838851,0.6666666666666667,0.41025641025641024,0.3589743589743589,0.12324841081988787,0.007165149729861791
40
+ ZC3HAV1+CEBPE,0.08024691358024691,0.08,0.08,0.08024691358024691,0.08024691358024691,0.15778688524590165,0.08,0.08,0.08,0.078,0.0006180401974148526,0.8148148148148148,0.8543173721436267,0.9506172839506173,162.0,976.0,0.20305246954480483,0.5139404519873891,0.09221914841694258,0.019781723262453504,0.0468182215665005,0.019781723262453504,0.0468182215665005,0.8461538461538461,0.8974358974358975,0.9487179487179487,0.12324841081988787,0.007165149729861791
transfer/code/CCFM/config/config_cascaded.py ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from dataclasses import dataclass
2
+ import os
3
+
4
+
5
+ @dataclass
6
+ class CascadedFlowConfig:
7
+ # === Base (same as scDFM FlowConfig) ===
8
+ model_type: str = "cascaded"
9
+ batch_size: int = 48
10
+ ntoken: int = 512
11
+ d_model: int = 128
12
+ nhead: int = 8
13
+ nlayers: int = 4
14
+ lr: float = 5e-5
15
+ steps: int = 200000
16
+ eta_min: float = 1e-6
17
+ devices: str = "1"
18
+ test_only: bool = False
19
+
20
+ data_name: str = "norman"
21
+ perturbation_function: str = "crisper"
22
+ noise_type: str = "Gaussian"
23
+ poisson_alpha: float = 0.8
24
+ poisson_target_sum: int = -1
25
+
26
+ print_every: int = 5000
27
+ mode: str = "predict_y"
28
+ result_path: str = "./result_online"
29
+ fusion_method: str = "differential_perceiver"
30
+ infer_top_gene: int = 1000
31
+ n_top_genes: int = 5000
32
+ checkpoint_path: str = ""
33
+ gamma: float = 0.5
34
+ split_method: str = "additive"
35
+ use_mmd_loss: bool = True
36
+ fold: int = 1
37
+ use_negative_edge: bool = True
38
+ topk: int = 30
39
+
40
+ # === Cascaded / Latent specific ===
41
+ scgpt_dim: int = 512
42
+ bottleneck_dim: int = 128
43
+ latent_weight: float = 1.0
44
+ choose_latent_p: float = 0.4
45
+ target_std: float = 1.0
46
+ dh_depth: int = 2
47
+ warmup_batches: int = 200 # batches to collect running stats
48
+
49
+ # === Cascaded noise (LatentForcing dino_first_cascaded_noised) ===
50
+ noise_beta: float = 0.25 # when training expr flow, t_latent ~ U[1-beta, 1] instead of 1.0
51
+
52
+ # === EMA ===
53
+ ema_decay: float = 0.9999
54
+
55
+ # === Logit-normal time-step sampling ===
56
+ t_sample_mode: str = "logit_normal" # "uniform" or "logit_normal"
57
+ t_expr_mean: float = 0.0 # expression flow logit-normal mu
58
+ t_expr_std: float = 1.0 # expression flow logit-normal sigma
59
+ t_latent_mean: float = 0.0 # latent flow logit-normal mu
60
+ t_latent_std: float = 1.0 # latent flow logit-normal sigma
61
+
62
+ # === LR warmup ===
63
+ warmup_steps: int = 2000
64
+
65
+ # === scGPT paths ===
66
+ scgpt_model_dir: str = "transfer/data/scGPT_pretrained"
67
+ scgpt_max_seq_len: int = 1200
68
+ scgpt_cache_path: str = "" # Pre-extracted HDF5 path. Empty = online extraction (default)
69
+
70
+ # === Inference ===
71
+ latent_steps: int = 20
72
+ expr_steps: int = 20
73
+ ode_method: str = "rk4" # "euler" or "rk4"
74
+
75
+ def __post_init__(self):
76
+ if self.data_name == "norman_umi_go_filtered":
77
+ self.n_top_genes = 5054
78
+ if self.data_name == "norman":
79
+ self.n_top_genes = 5000
80
+
81
+ def make_path(self):
82
+ scgpt_mode = "cached" if self.scgpt_cache_path else "online"
83
+ t_mode = "ln" if self.t_sample_mode == "logit_normal" else "uni"
84
+ exp_name = (
85
+ f"ccfm-{self.data_name}-f{self.fold}"
86
+ f"-topk{self.topk}-neg{self.use_negative_edge}"
87
+ f"-d{self.d_model}-lr{self.lr}"
88
+ f"-lw{self.latent_weight}-lp{self.choose_latent_p}"
89
+ f"-ema{self.ema_decay}-{t_mode}-wu{self.warmup_steps}"
90
+ f"-{self.ode_method}-{scgpt_mode}"
91
+ )
92
+ return os.path.join(self.result_path, exp_name)
transfer/code/CCFM/config/config_cascaded.py.bak ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from dataclasses import dataclass
2
+ import os
3
+
4
+
5
+ @dataclass
6
+ class CascadedFlowConfig:
7
+ # === Base (same as scDFM FlowConfig) ===
8
+ model_type: str = "cascaded"
9
+ batch_size: int = 48
10
+ ntoken: int = 512
11
+ d_model: int = 128
12
+ nhead: int = 8
13
+ nlayers: int = 4
14
+ lr: float = 5e-5
15
+ steps: int = 200000
16
+ eta_min: float = 1e-6
17
+ devices: str = "1"
18
+ test_only: bool = False
19
+
20
+ data_name: str = "norman"
21
+ perturbation_function: str = "crisper"
22
+ noise_type: str = "Gaussian"
23
+ poisson_alpha: float = 0.8
24
+ poisson_target_sum: int = -1
25
+
26
+ print_every: int = 5000
27
+ mode: str = "predict_y"
28
+ result_path: str = "./result_online"
29
+ fusion_method: str = "differential_perceiver"
30
+ infer_top_gene: int = 1000
31
+ n_top_genes: int = 5000
32
+ checkpoint_path: str = ""
33
+ gamma: float = 0.5
34
+ split_method: str = "additive"
35
+ use_mmd_loss: bool = True
36
+ fold: int = 1
37
+ use_negative_edge: bool = True
38
+ topk: int = 30
39
+
40
+ # === Cascaded / Latent specific ===
41
+ scgpt_dim: int = 512
42
+ bottleneck_dim: int = 128
43
+ latent_weight: float = 1.0
44
+ choose_latent_p: float = 0.4
45
+ target_std: float = 1.0
46
+ dh_depth: int = 2
47
+ warmup_batches: int = 200 # batches to collect running stats
48
+
49
+ # === EMA ===
50
+ ema_decay: float = 0.9999
51
+
52
+ # === Logit-normal time-step sampling ===
53
+ t_sample_mode: str = "logit_normal" # "uniform" or "logit_normal"
54
+ t_expr_mean: float = 0.0 # expression flow logit-normal mu
55
+ t_expr_std: float = 1.0 # expression flow logit-normal sigma
56
+ t_latent_mean: float = 0.0 # latent flow logit-normal mu
57
+ t_latent_std: float = 1.0 # latent flow logit-normal sigma
58
+
59
+ # === LR warmup ===
60
+ warmup_steps: int = 2000
61
+
62
+ # === scGPT paths ===
63
+ scgpt_model_dir: str = "transfer/data/scGPT_pretrained"
64
+ scgpt_max_seq_len: int = 1200
65
+ scgpt_cache_path: str = "" # Pre-extracted HDF5 path. Empty = online extraction (default)
66
+
67
+ # === Inference ===
68
+ latent_steps: int = 20
69
+ expr_steps: int = 20
70
+ ode_method: str = "rk4" # "euler" or "rk4"
71
+
72
+ def __post_init__(self):
73
+ if self.data_name == "norman_umi_go_filtered":
74
+ self.n_top_genes = 5054
75
+ if self.data_name == "norman":
76
+ self.n_top_genes = 5000
77
+
78
+ def make_path(self):
79
+ scgpt_mode = "cached" if self.scgpt_cache_path else "online"
80
+ t_mode = "ln" if self.t_sample_mode == "logit_normal" else "uni"
81
+ exp_name = (
82
+ f"ccfm-{self.data_name}-f{self.fold}"
83
+ f"-topk{self.topk}-neg{self.use_negative_edge}"
84
+ f"-d{self.d_model}-lr{self.lr}"
85
+ f"-lw{self.latent_weight}-lp{self.choose_latent_p}"
86
+ f"-ema{self.ema_decay}-{t_mode}-wu{self.warmup_steps}"
87
+ f"-{self.ode_method}-{scgpt_mode}"
88
+ )
89
+ return os.path.join(self.result_path, exp_name)
transfer/code/CCFM/eval_ccfm_v2.sh ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/sh
2
+ #PJM -L rscgrp=b-batch
3
+ #PJM -L gpu=1
4
+ #PJM -L elapse=3:00:00
5
+ #PJM -N eval_ccfm_v2
6
+ #PJM -j
7
+ #PJM -o logs/eval_ccfm_v2_%j.out
8
+
9
+ module load cuda/12.2.2
10
+ module load cudnn/8.9.7
11
+ module load gcc-toolset/12
12
+
13
+ source /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/activate
14
+ cd /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM
15
+
16
+ export PYTORCH_CUDA_ALLOC_CONF=max_split_size_mb:256
17
+
18
+ CKPT="/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_100000/checkpoint.pt"
19
+
20
+ echo "=========================================="
21
+ echo "Job ID: $PJM_JOBID"
22
+ echo "Job Name: $PJM_JOBNAME"
23
+ echo "Start: $(date)"
24
+ echo "Node: $(hostname)"
25
+ echo "GPU: $(nvidia-smi --query-gpu=name,memory.total --format=csv,noheader 2>/dev/null || echo 'N/A')"
26
+ echo "Eval: CCFM v2 (online) iteration_100000"
27
+ echo "=========================================="
28
+
29
+ python scripts/run_cascaded.py \
30
+ --data-name norman \
31
+ --d-model 128 \
32
+ --nhead 8 \
33
+ --nlayers 4 \
34
+ --batch-size 48 \
35
+ --lr 5e-5 \
36
+ --steps 200000 \
37
+ --fusion-method differential_perceiver \
38
+ --perturbation-function crisper \
39
+ --noise-type Gaussian \
40
+ --infer-top-gene 1000 \
41
+ --n-top-genes 5000 \
42
+ --use-mmd-loss \
43
+ --gamma 0.5 \
44
+ --split-method additive \
45
+ --fold 1 \
46
+ --scgpt-dim 512 \
47
+ --bottleneck-dim 128 \
48
+ --latent-weight 1.0 \
49
+ --choose-latent-p 0.4 \
50
+ --dh-depth 2 \
51
+ --print-every 10000 \
52
+ --topk 30 \
53
+ --use-negative-edge \
54
+ --ema-decay 0.9999 \
55
+ --t-sample-mode logit_normal \
56
+ --t-expr-mean 0.0 \
57
+ --t-expr-std 1.0 \
58
+ --t-latent-mean 0.0 \
59
+ --t-latent-std 1.0 \
60
+ --warmup-steps 2000 \
61
+ --ode-method rk4 \
62
+ --result-path ./result \
63
+ --checkpoint-path "$CKPT" \
64
+ --test-only
65
+
66
+ echo "=========================================="
67
+ echo "Finished: $(date)"
68
+ echo "=========================================="
transfer/code/CCFM/eval_joint_generate.sh ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ #PJM -L rscgrp=b-batch
3
+ #PJM -L gpu=1
4
+ #PJM -L elapse=3:00:00
5
+ #PJM -j
6
+ #PJM -S
7
+ #PJM -o /home/hp250092/ku50001222/qian/aivc/lfj/transfer/logs/ccfm_eval_joint_%j.out
8
+
9
+ # Evaluate CCFM with joint-generation inference (LatentForcing style)
10
+ # using the checkpoint from two-stage ODE training.
11
+
12
+ source /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/activate
13
+ cd /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM
14
+
15
+ export PYTORCH_CUDA_ALLOC_CONF=max_split_size_mb:256
16
+
17
+ echo "=========================================="
18
+ echo "Job ID: $PJM_JOBID"
19
+ echo "Start: $(date)"
20
+ echo "Node: $(hostname)"
21
+ echo "Eval mode: joint-generate (LatentForcing style)"
22
+ echo "=========================================="
23
+
24
+ CHECKPOINT="/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/checkpoint.pt"
25
+
26
+ python scripts/run_cascaded.py \
27
+ --data-name norman \
28
+ --d-model 128 \
29
+ --nhead 8 \
30
+ --nlayers 4 \
31
+ --batch-size 128 \
32
+ --lr 5e-5 \
33
+ --steps 200000 \
34
+ --fusion-method differential_perceiver \
35
+ --perturbation-function crisper \
36
+ --noise-type Gaussian \
37
+ --infer-top-gene 1000 \
38
+ --n-top-genes 5000 \
39
+ --use-mmd-loss \
40
+ --gamma 0.5 \
41
+ --split-method additive \
42
+ --fold 1 \
43
+ --scgpt-dim 512 \
44
+ --bottleneck-dim 128 \
45
+ --latent-weight 1.0 \
46
+ --choose-latent-p 0.4 \
47
+ --dh-depth 2 \
48
+ --latent-steps 20 \
49
+ --expr-steps 20 \
50
+ --result-path ./result_joint_generate \
51
+ --checkpoint-path "$CHECKPOINT" \
52
+ --test-only
53
+
54
+ echo "=========================================="
55
+ echo "Finished: $(date)"
56
+ echo "=========================================="
transfer/code/CCFM/eval_joint_generate.sh.5404217.stats ADDED
@@ -0,0 +1,467 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Job Statistical Information
2
+
3
+ JOB ID : 5404217
4
+ SUB JOB NUM : -
5
+ START BULKNO : -
6
+ END BULKNO : -
7
+ HOST NAME : genkai0001
8
+ JOB NAME : eval_joint_generate.sh
9
+ JOB TYPE : BATCH
10
+ JOB MODEL : NM
11
+ USER : ku50001222
12
+ GROUP : hp250092
13
+ RESOURCE UNIT : rscunit_pg01 <DEFAULT>
14
+ RESOURCE GROUP : b-batch
15
+ APRIORITY : 127 <DEFAULT>
16
+ PRIORITY : 127 <DEFAULT>
17
+ SHELL : /bin/bash
18
+ MAIL SEND FLAG : ----
19
+ MAIL ADDRESS : ku50001222
20
+ STEP DEPENDENCY EXP :
21
+ FILE MASK : 0022
22
+ STANDARD OUT FILE : /home/hp250092/ku50001222/qian/aivc/lfj/transfer/logs/ccfm_eval_joint_5404217.out
23
+ STANDARD ERR FILE : /home/hp250092/ku50001222/qian/aivc/lfj/transfer/logs/ccfm_eval_joint_5404217.out
24
+ INFORMATION FILE : /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/eval_joint_generate.sh.5404217.stats
25
+ PJSUB DIRECTORY : /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM
26
+ FILE SYSTEM NAME :
27
+ APPLICATION NAME :
28
+ NUMA POLICY : 0
29
+ SPECIFIED JOB START DATE : -
30
+ AUTO RESTART : yes
31
+ MPI PROC (REQUIRE) : 1
32
+ MPI PROC (ALLOC) : 1
33
+ MPI STATIC PROC (REQUIRE) : 1
34
+ MPI STATIC PROC (ALLOC) : 1
35
+ PROC CPU LIMIT : unlimited <DEFAULT>
36
+ PROC COREFILE LIMIT : 0.0 MB (0) <DEFAULT>
37
+ PROC CREATE PROCESS LIMIT : 16384 <DEFAULT>
38
+ PROC DATA LIMIT : unlimited <DEFAULT>
39
+ PROC LOCKED MEMORY LIMIT : unlimited <DEFAULT>
40
+ PROC MESSAGE QUEUE LIMIT : unlimited <DEFAULT>
41
+ PROC OPEN FILES LIMIT : 1048576 <DEFAULT>
42
+ PROC PENDING SIGNAL LIMIT : unlimited <DEFAULT>
43
+ PROC PERMFILE LIMIT : unlimited <DEFAULT>
44
+ PROC STACK LIMIT : unlimited <DEFAULT>
45
+ PROC VIRTUAL MEMORY LIMIT : unlimited <DEFAULT>
46
+ COMMENT :
47
+ SUPPLEMENTARY INFORMATION :
48
+ RETRY NUM : 0
49
+ LAST STATE : RNO
50
+ STATE : EXT
51
+ BACKFILL : 1
52
+ ACCEPT DATE : 2026/03/14 20:43:33
53
+ QUEUED DATE : 2026/03/14 20:43:33
54
+ PRO START DATE : -
55
+ PRO END DATE : -
56
+ JOB START DATE : 2026/03/14 23:27:39
57
+ JOB END DATE : 2026/03/14 23:38:49
58
+ EPI START DATE : -
59
+ EPI END DATE : -
60
+ EXIT DATE : 2026/03/14 23:38:50
61
+ JOB DELETE DATE (REQUIRE) : -
62
+ JOB DELETE DATE : -
63
+ ALL PREC SUBJOB EXIT DATE : -
64
+ HOLD TIME : 0
65
+ LAST HOLD USER : -
66
+ HOLD NUM : 0
67
+ PRM DATE : 2026/03/14 23:38:49
68
+ AFFECTED NODE ID : -
69
+ EXIT CODE : 0
70
+ SIGNAL NO : -
71
+ PJM CODE : 0
72
+ REASON : -
73
+ PRO EXIT CODE : -
74
+ EPI EXIT CODE : -
75
+ ELAPSE TIME (LIMIT) : 03:00:00 (10800)
76
+ ELAPSE TIME (USE) : 00:11:10 (670)
77
+ USER CPU TIME (USE) : 717260 ms
78
+ SYSTEM CPU TIME (USE) : 16400 ms
79
+ CPU TIME (TOTAL) : 733660 ms
80
+ MEMORY SIZE (LIMIT) : -
81
+ MEMORY SIZE (ALLOC) : -
82
+ MAX MEMORY SIZE (USE) : -
83
+ VNODE MEMORY SIZE (LIMIT) : 232050.0 MiB (243322060800)
84
+ VNODE MEMORY SIZE (ALLOC) : 232050.0 MiB (243322060800)
85
+ VNODE MAX MEMORY SIZE (USE) : 8343.9 MiB (8749178880)
86
+ CPU NUM (REQUIRE) : 30
87
+ CPU NUM (ALLOC) : 30
88
+ CPU NUM (USE) : 30
89
+ TOFU USER COMM RECV BYTE : -
90
+ TOFU USER COMM SEND BYTE : -
91
+ TOFU SYSTEM COMM RECV BYTE : -
92
+ TOFU SYSTEM COMM SEND BYTE : -
93
+ FJ PROFILER : -
94
+ SECTOR CACHE : -
95
+ INTRA NODE BARRIER : -
96
+ AVG POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
97
+ MAX POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
98
+ MIN POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
99
+ ENERGY CONSUMPTION OF CORES/CMG (IDEAL) : -
100
+ AVG POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
101
+ MAX POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
102
+ MIN POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
103
+ ENERGY CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
104
+ AVG POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
105
+ MAX POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
106
+ MIN POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
107
+ ENERGY CONSUMPTION OF MEM/CMG (IDEAL) : -
108
+ AVG POWER CONSUMPTION OF TOFU (IDEAL) : -
109
+ MAX POWER CONSUMPTION OF TOFU (IDEAL) : -
110
+ MIN POWER CONSUMPTION OF TOFU (IDEAL) : -
111
+ ENERGY CONSUMPTION OF TOFU (IDEAL) : -
112
+ AVG POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
113
+ MAX POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
114
+ MIN POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
115
+ ENERGY CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
116
+ AVG POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
117
+ MAX POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
118
+ MIN POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
119
+ ENERGY CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
120
+ AVG POWER CONSUMPTION OF PCI-E (IDEAL) : -
121
+ MAX POWER CONSUMPTION OF PCI-E (IDEAL) : -
122
+ MIN POWER CONSUMPTION OF PCI-E (IDEAL) : -
123
+ ENERGY CONSUMPTION OF PCI-E (IDEAL) : -
124
+ AVG POWER CONSUMPTION OF NODE (IDEAL) : -
125
+ MAX POWER CONSUMPTION OF NODE (IDEAL) : -
126
+ MIN POWER CONSUMPTION OF NODE (IDEAL) : -
127
+ ENERGY CONSUMPTION OF NODE (IDEAL) : -
128
+ AVG POWER CONSUMPTION OF NODE (MEASURED) : -
129
+ MAX POWER CONSUMPTION OF NODE (MEASURED) : -
130
+ MIN POWER CONSUMPTION OF NODE (MEASURED) : -
131
+ ENERGY CONSUMPTION OF NODE (MEASURED) : -
132
+ UTILIZATION INFO OF POWER API : 0
133
+ AVG POWER CONSUMPTION OF CPU/PKG : -
134
+ MAX POWER CONSUMPTION OF CPU/PKG : -
135
+ MIN POWER CONSUMPTION OF CPU/PKG : -
136
+ ENERGY CONSUMPTION OF CPU/PKG : -
137
+ AVG POWER CONSUMPTION OF MEM/PKG : -
138
+ MAX POWER CONSUMPTION OF MEM/PKG : -
139
+ MIN POWER CONSUMPTION OF MEM/PKG : -
140
+ ENERGY CONSUMPTION OF MEM/PKG : -
141
+ AVG POWER CONSUMPTION OF PP0/PKG : -
142
+ MAX POWER CONSUMPTION OF PP0/PKG : -
143
+ MIN POWER CONSUMPTION OF PP0/PKG : -
144
+ ENERGY CONSUMPTION OF PP0/PKG : -
145
+ POWER CONSUMPTION STATE : 4
146
+ POWER CONSUMPTION MEASURE START DATE : -
147
+ POWER CONSUMPTION MEASURE END DATE : -
148
+ VNODE NUM (REQUIRE) : 1
149
+ VNODE NUM (ALLOC) : 1
150
+ VNODE NUM (USE) : 1
151
+ NODE NUM (REQUIRE) : -
152
+ NODE NUM (ALLOC) : -
153
+ NODE NUM (USE) : -
154
+ NODE NUM (UNUSED) : 0
155
+ NODE ID (USE) : 0x02FF001C
156
+ TOFU COORDINATE (USE) : -
157
+ NODE ID (RANK) : 0x02FF001C(0)
158
+ jobenv (REQUIRE) :
159
+ jobenv (ALLOC) :
160
+ jobenv (USE) :
161
+ gpu (REQUIRE) : 1
162
+ gpu (ALLOC) : 1
163
+ gpu (USE) : -
164
+ simplex (REQUIRE) :
165
+ simplex (ALLOC) :
166
+ simplex (USE) :
167
+ shared (REQUIRE) : true
168
+ shared (ALLOC) : true
169
+ shared (USE) :
170
+ short-job (REQUIRE) : false
171
+ short-job (ALLOC) : false
172
+ short-job (USE) :
173
+ rsc001 (REQUIRE) : 0 <DEFAULT>
174
+ rsc001 (ALLOC) : 0
175
+ rsc001 (USE) : -
176
+ rsc002 (REQUIRE) : 0 <DEFAULT>
177
+ rsc002 (ALLOC) : 0
178
+ rsc002 (USE) : -
179
+ rsc003 (REQUIRE) : 0 <DEFAULT>
180
+ rsc003 (ALLOC) : 0
181
+ rsc003 (USE) : -
182
+ rsc004 (REQUIRE) : 0 <DEFAULT>
183
+ rsc004 (ALLOC) : 0
184
+ rsc004 (USE) : -
185
+ rsc005 (REQUIRE) : 0 <DEFAULT>
186
+ rsc005 (ALLOC) : 0
187
+ rsc005 (USE) : -
188
+ rsc006 (REQUIRE) : 0 <DEFAULT>
189
+ rsc006 (ALLOC) : 0
190
+ rsc006 (USE) : -
191
+ rsc007 (REQUIRE) : 0 <DEFAULT>
192
+ rsc007 (ALLOC) : 0
193
+ rsc007 (USE) : -
194
+ rsc008 (REQUIRE) : 0 <DEFAULT>
195
+ rsc008 (ALLOC) : 0
196
+ rsc008 (USE) : -
197
+ rsc009 (REQUIRE) : 0 <DEFAULT>
198
+ rsc009 (ALLOC) : 0
199
+ rsc009 (USE) : -
200
+ rsc010 (REQUIRE) : 0 <DEFAULT>
201
+ rsc010 (ALLOC) : 0
202
+ rsc010 (USE) : -
203
+ rsc011 (REQUIRE) : 0 <DEFAULT>
204
+ rsc011 (ALLOC) : 0
205
+ rsc011 (USE) : -
206
+ rsc012 (REQUIRE) : 0 <DEFAULT>
207
+ rsc012 (ALLOC) : 0
208
+ rsc012 (USE) : -
209
+ rsc013 (REQUIRE) : 0 <DEFAULT>
210
+ rsc013 (ALLOC) : 0
211
+ rsc013 (USE) : -
212
+ rsc014 (REQUIRE) : 0 <DEFAULT>
213
+ rsc014 (ALLOC) : 0
214
+ rsc014 (USE) : -
215
+ rsc015 (REQUIRE) : 0 <DEFAULT>
216
+ rsc015 (ALLOC) : 0
217
+ rsc015 (USE) : -
218
+ rsc016 (REQUIRE) : 0 <DEFAULT>
219
+ rsc016 (ALLOC) : 0
220
+ rsc016 (USE) : -
221
+ rsc017 (REQUIRE) : 0 <DEFAULT>
222
+ rsc017 (ALLOC) : 0
223
+ rsc017 (USE) : -
224
+ rsc018 (REQUIRE) : 0 <DEFAULT>
225
+ rsc018 (ALLOC) : 0
226
+ rsc018 (USE) : -
227
+ rsc019 (REQUIRE) : 0 <DEFAULT>
228
+ rsc019 (ALLOC) : 0
229
+ rsc019 (USE) : -
230
+ rsc020 (REQUIRE) : 0 <DEFAULT>
231
+ rsc020 (ALLOC) : 0
232
+ rsc020 (USE) : -
233
+ rsc021 (REQUIRE) : 0 <DEFAULT>
234
+ rsc021 (ALLOC) : 0
235
+ rsc021 (USE) : -
236
+ rsc022 (REQUIRE) : 0 <DEFAULT>
237
+ rsc022 (ALLOC) : 0
238
+ rsc022 (USE) : -
239
+ rsc023 (REQUIRE) : 0 <DEFAULT>
240
+ rsc023 (ALLOC) : 0
241
+ rsc023 (USE) : -
242
+ rsc024 (REQUIRE) : 0 <DEFAULT>
243
+ rsc024 (ALLOC) : 0
244
+ rsc024 (USE) : -
245
+ rsc025 (REQUIRE) : 0 <DEFAULT>
246
+ rsc025 (ALLOC) : 0
247
+ rsc025 (USE) : -
248
+ rsc026 (REQUIRE) : 0 <DEFAULT>
249
+ rsc026 (ALLOC) : 0
250
+ rsc026 (USE) : -
251
+ rsc027 (REQUIRE) : 0 <DEFAULT>
252
+ rsc027 (ALLOC) : 0
253
+ rsc027 (USE) : -
254
+ rsc028 (REQUIRE) : 0 <DEFAULT>
255
+ rsc028 (ALLOC) : 0
256
+ rsc028 (USE) : -
257
+ rsc029 (REQUIRE) : 0 <DEFAULT>
258
+ rsc029 (ALLOC) : 0
259
+ rsc029 (USE) : -
260
+ rsc030 (REQUIRE) : 0 <DEFAULT>
261
+ rsc030 (ALLOC) : 0
262
+ rsc030 (USE) : -
263
+ rsc031 (REQUIRE) : 0 <DEFAULT>
264
+ rsc031 (ALLOC) : 0
265
+ rsc031 (USE) : -
266
+ rsc032 (REQUIRE) : 0 <DEFAULT>
267
+ rsc032 (ALLOC) : 0
268
+ rsc032 (USE) : -
269
+ rsc033 (REQUIRE) : 0 <DEFAULT>
270
+ rsc033 (ALLOC) : 0
271
+ rsc033 (USE) : -
272
+ rsc034 (REQUIRE) : 0 <DEFAULT>
273
+ rsc034 (ALLOC) : 0
274
+ rsc034 (USE) : -
275
+ rsc035 (REQUIRE) : 0 <DEFAULT>
276
+ rsc035 (ALLOC) : 0
277
+ rsc035 (USE) : -
278
+ rsc036 (REQUIRE) : 0 <DEFAULT>
279
+ rsc036 (ALLOC) : 0
280
+ rsc036 (USE) : -
281
+ rsc037 (REQUIRE) : 0 <DEFAULT>
282
+ rsc037 (ALLOC) : 0
283
+ rsc037 (USE) : -
284
+ rsc038 (REQUIRE) : 0 <DEFAULT>
285
+ rsc038 (ALLOC) : 0
286
+ rsc038 (USE) : -
287
+ rsc039 (REQUIRE) : 0 <DEFAULT>
288
+ rsc039 (ALLOC) : 0
289
+ rsc039 (USE) : -
290
+ rsc040 (REQUIRE) : 0 <DEFAULT>
291
+ rsc040 (ALLOC) : 0
292
+ rsc040 (USE) : -
293
+ rsc041 (REQUIRE) : 0 <DEFAULT>
294
+ rsc041 (ALLOC) : 0
295
+ rsc041 (USE) : -
296
+ rsc042 (REQUIRE) : 0 <DEFAULT>
297
+ rsc042 (ALLOC) : 0
298
+ rsc042 (USE) : -
299
+ rsc043 (REQUIRE) : 0 <DEFAULT>
300
+ rsc043 (ALLOC) : 0
301
+ rsc043 (USE) : -
302
+ rsc044 (REQUIRE) : 0 <DEFAULT>
303
+ rsc044 (ALLOC) : 0
304
+ rsc044 (USE) : -
305
+ rsc045 (REQUIRE) : 0 <DEFAULT>
306
+ rsc045 (ALLOC) : 0
307
+ rsc045 (USE) : -
308
+ rsc046 (REQUIRE) : 0 <DEFAULT>
309
+ rsc046 (ALLOC) : 0
310
+ rsc046 (USE) : -
311
+ rsc047 (REQUIRE) : 0 <DEFAULT>
312
+ rsc047 (ALLOC) : 0
313
+ rsc047 (USE) : -
314
+ rsc048 (REQUIRE) : 0 <DEFAULT>
315
+ rsc048 (ALLOC) : 0
316
+ rsc048 (USE) : -
317
+ rsc049 (REQUIRE) : 0 <DEFAULT>
318
+ rsc049 (ALLOC) : 0
319
+ rsc049 (USE) : -
320
+ rsc050 (REQUIRE) : 0 <DEFAULT>
321
+ rsc050 (ALLOC) : 0
322
+ rsc050 (USE) : -
323
+
324
+ --------------------------------------------------------------------------------
325
+ Node Statistical Information
326
+
327
+ VNODE ID : 0
328
+ NODE ID : 0x02FF001C
329
+ TOFU COORDINATE : -
330
+ NODE COORDINATE : -
331
+ RANK NO : 0
332
+ NODE AVAILABLE : available
333
+ ARCHI INFORMATION : PG
334
+ CPU BITMAP (ALLOC) : 0xfffffffc000000
335
+ CPU BITMAP (USE) : 0xfffffffc000000
336
+ MPI PROC (ALLOC) : 1
337
+ MPI STATIC PROC (ALLOC) : 1
338
+ USER CPU TIME (USE) : 717260 ms
339
+ SYSTEM CPU TIME (USE) : 16400 ms
340
+ CPU TIME (TOTAL) : 733660 ms
341
+ MEMORY SIZE (ALLOC) : 232050.0 MiB (243322060800)
342
+ MAX MEMORY SIZE (USE) : 8343.9 MiB (8749178880)
343
+ CPU NUM (ALLOC) : 30
344
+ CPU NUM (USE) : 30
345
+ PERF COUNT 1 : -
346
+ PERF COUNT 2 : -
347
+ PERF COUNT 3 : -
348
+ PERF COUNT 4 : -
349
+ PERF COUNT 5 : -
350
+ PERF COUNT 6 : -
351
+ PERF COUNT 7 : -
352
+ PERF COUNT 8 : -
353
+ PERF COUNT 9 : -
354
+ TOFU USER COMM RECV BYTE : -
355
+ TOFU USER COMM SEND BYTE : -
356
+ TOFU SYSTEM COMM RECV BYTE : -
357
+ TOFU SYSTEM COMM SEND BYTE : -
358
+ FJ PROFILER : -
359
+ SECTOR CACHE : -
360
+ INTRA NODE BARRIER : -
361
+ AVG POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
362
+ MAX POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
363
+ MIN POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
364
+ ENERGY CONSUMPTION OF CORES/CMG (IDEAL) : -
365
+ AVG POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
366
+ MAX POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
367
+ MIN POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
368
+ ENERGY CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
369
+ AVG POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
370
+ MAX POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
371
+ MIN POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
372
+ ENERGY CONSUMPTION OF MEM/CMG (IDEAL) : -
373
+ AVG POWER CONSUMPTION OF TOFU (IDEAL) : -
374
+ MAX POWER CONSUMPTION OF TOFU (IDEAL) : -
375
+ MIN POWER CONSUMPTION OF TOFU (IDEAL) : -
376
+ ENERGY CONSUMPTION OF TOFU (IDEAL) : -
377
+ AVG POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
378
+ MAX POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
379
+ MIN POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
380
+ ENERGY CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
381
+ AVG POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
382
+ MAX POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
383
+ MIN POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
384
+ ENERGY CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
385
+ AVG POWER CONSUMPTION OF PCI-E (IDEAL) : -
386
+ MAX POWER CONSUMPTION OF PCI-E (IDEAL) : -
387
+ MIN POWER CONSUMPTION OF PCI-E (IDEAL) : -
388
+ ENERGY CONSUMPTION OF PCI-E (IDEAL) : -
389
+ AVG POWER CONSUMPTION OF NODE (IDEAL) : -
390
+ MAX POWER CONSUMPTION OF NODE (IDEAL) : -
391
+ MIN POWER CONSUMPTION OF NODE (IDEAL) : -
392
+ ENERGY CONSUMPTION OF NODE (IDEAL) : -
393
+ AVG POWER CONSUMPTION OF NODE (MEASURED) : -
394
+ MAX POWER CONSUMPTION OF NODE (MEASURED) : -
395
+ MIN POWER CONSUMPTION OF NODE (MEASURED) : -
396
+ ENERGY CONSUMPTION OF NODE (MEASURED) : -
397
+ UTILIZATION INFO OF POWER API : 0
398
+ AVG POWER CONSUMPTION OF CPU/PKG : -
399
+ MAX POWER CONSUMPTION OF CPU/PKG : -
400
+ MIN POWER CONSUMPTION OF CPU/PKG : -
401
+ ENERGY CONSUMPTION OF CPU/PKG : -
402
+ AVG POWER CONSUMPTION OF MEM/PKG : -
403
+ MAX POWER CONSUMPTION OF MEM/PKG : -
404
+ MIN POWER CONSUMPTION OF MEM/PKG : -
405
+ ENERGY CONSUMPTION OF MEM/PKG : -
406
+ AVG POWER CONSUMPTION OF PP0/PKG : -
407
+ MAX POWER CONSUMPTION OF PP0/PKG : -
408
+ MIN POWER CONSUMPTION OF PP0/PKG : -
409
+ ENERGY CONSUMPTION OF PP0/PKG : -
410
+ POWER CONSUMPTION STATE : 4
411
+ POWER CONSUMPTION MEASURE START DATE : -
412
+ POWER CONSUMPTION MEASURE END DATE : -
413
+ gpu (REQUIRE) : 1
414
+ gpu (ALLOC) : 1
415
+ gpu (USE) : -
416
+ simplex (REQUIRE) :
417
+ simplex (ALLOC) :
418
+ simplex (USE) :
419
+ shared (REQUIRE) : true
420
+ shared (ALLOC) : true
421
+ shared (USE) :
422
+ short-job (REQUIRE) : false
423
+ short-job (ALLOC) : false
424
+ short-job (USE) :
425
+ rsc001 (REQUIRE) : 0 <DEFAULT>
426
+ rsc001 (ALLOC) : 0
427
+ rsc001 (USE) : -
428
+ rsc002 (REQUIRE) : 0 <DEFAULT>
429
+ rsc002 (ALLOC) : 0
430
+ rsc002 (USE) : -
431
+ rsc003 (REQUIRE) : 0 <DEFAULT>
432
+ rsc003 (ALLOC) : 0
433
+ rsc003 (USE) : -
434
+ rsc004 (REQUIRE) : 0 <DEFAULT>
435
+ rsc004 (ALLOC) : 0
436
+ rsc004 (USE) : -
437
+ rsc005 (REQUIRE) : 0 <DEFAULT>
438
+ rsc005 (ALLOC) : 0
439
+ rsc005 (USE) : -
440
+ rsc006 (REQUIRE) : 0 <DEFAULT>
441
+ rsc006 (ALLOC) : 0
442
+ rsc006 (USE) : -
443
+ rsc007 (REQUIRE) : 0 <DEFAULT>
444
+ rsc007 (ALLOC) : 0
445
+ rsc007 (USE) : -
446
+ rsc008 (REQUIRE) : 0 <DEFAULT>
447
+ rsc008 (ALLOC) : 0
448
+ rsc008 (USE) : -
449
+ rsc009 (REQUIRE) : 0 <DEFAULT>
450
+ rsc009 (ALLOC) : 0
451
+ rsc009 (USE) : -
452
+ rsc010 (REQUIRE) : 0 <DEFAULT>
453
+ rsc010 (ALLOC) : 0
454
+ rsc010 (USE) : -
455
+ rsc011 (REQUIRE) : 0 <DEFAULT>
456
+ rsc011 (ALLOC) : 0
457
+ rsc011 (USE) : -
458
+ rsc012 (REQUIRE) : 0 <DEFAULT>
459
+ rsc012 (ALLOC) : 0
460
+ rsc012 (USE) : -
461
+ rsc013 (REQUIRE) : 0 <DEFAULT>
462
+ rsc013 (ALLOC) : 0
463
+ rsc013 (USE) : -
464
+ rsc014 (REQUIRE) : 0 <DEFAULT>
465
+ rsc014 (ALLOC) : 0
466
+ rsc014 (USE) : -
467
+
transfer/code/CCFM/logs/ccfm_1gpu_cached_5404438.out ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/200000 [00:00<?, ?it/s]Traceback (most recent call last):
 
 
 
 
 
 
 
1
  0%| | 0/200000 [00:42<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==========================================
2
+ Job ID: 5404438
3
+ Job Name: ccfm_1gpu_cached
4
+ Start: Sun Mar 15 00:10:32 JST 2026
5
+ Node: b0031
6
+ GPU: NVIDIA H100, 95830 MiB
7
+ Run: CCFM 1GPU cached scGPT
8
+ ==========================================
9
+ The following values were not passed to `accelerate launch` and had defaults used instead:
10
+ `--num_machines` was set to a value of `1`
11
+ `--mixed_precision` was set to a value of `'no'`
12
+ `--dynamo_backend` was set to a value of `'no'`
13
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
14
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
15
+ warnings.warn("flash_attn is not installed")
16
+ WARNING:accelerate.utils.other:[RANK 0] Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
17
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
18
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
19
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
20
+ adata.uns["hvg"] = {"flavor": flavor}
21
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
22
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
23
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
24
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
25
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
26
+ warnings.warn(
27
+ CascadedFlowConfig(model_type='cascaded', batch_size=48, ntoken=512, d_model=128, nhead=8, nlayers=4, lr=5e-05, steps=200000, eta_min=1e-06, devices='1', test_only=False, data_name='norman', perturbation_function='crisper', noise_type='Gaussian', poisson_alpha=0.8, poisson_target_sum=-1, print_every=5000, mode='predict_y', result_path='./result_scgpt', fusion_method='differential_perceiver', infer_top_gene=1000, n_top_genes=5000, checkpoint_path='', gamma=0.5, split_method='additive', use_mmd_loss=True, fold=1, use_negative_edge=True, topk=30, scgpt_dim=512, bottleneck_dim=128, latent_weight=1.0, choose_latent_p=0.4, target_std=1.0, dh_depth=2, warmup_batches=200, scgpt_model_dir='transfer/data/scGPT_pretrained', scgpt_max_seq_len=1200, scgpt_cache_path='scgpt_cache_norman.h5', latent_steps=20, expr_steps=20)
28
+ Converted var_names to gene symbols, sample: ['RP11-34P13.8', 'RP11-54O7.3', 'SAMD11', 'PERM1', 'HES4']
29
+ Warning: ctrl is not in the gene names
30
+ ##### loading vocab from file #####
31
+ Loaded 159/163 pretrained parameters
32
+ Using pre-extracted scGPT cache: scgpt_cache_norman.h5
33
+ Cache shape: (91205, 5035, 512), cells: 91205
34
+
35
  0%| | 0/200000 [00:00<?, ?it/s]Traceback (most recent call last):
36
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 324, in <module>
37
+ cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
38
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 50, in lookup
39
+ row_indices = [self.name_to_idx[n] for n in cell_names]
40
+ ~~~~~~~~~~~~~~~~^^^
41
+ KeyError: ('TGAGCCGCAGTAAGAT-8',)
42
+
43
  0%| | 0/200000 [00:42<?, ?it/s]
44
+ Traceback (most recent call last):
45
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/accelerate", line 8, in <module>
46
+ sys.exit(main())
47
+ ~~~~^^
48
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/accelerate_cli.py", line 50, in main
49
+ args.func(args)
50
+ ~~~~~~~~~^^^^^^
51
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1405, in launch_command
52
+ simple_launcher(args)
53
+ ~~~~~~~~~~~~~~~^^^^^^
54
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 993, in simple_launcher
55
+ raise subprocess.CalledProcessError(returncode=process.returncode, cmd=cmd)
56
+ subprocess.CalledProcessError: Command '['/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/python', 'scripts/run_cascaded.py', '--data-name', 'norman', '--d-model', '128', '--nhead', '8', '--nlayers', '4', '--batch-size', '48', '--lr', '5e-5', '--steps', '200000', '--fusion-method', 'differential_perceiver', '--perturbation-function', 'crisper', '--noise-type', 'Gaussian', '--infer-top-gene', '1000', '--n-top-genes', '5000', '--use-mmd-loss', '--gamma', '0.5', '--split-method', 'additive', '--fold', '1', '--scgpt-dim', '512', '--bottleneck-dim', '128', '--latent-weight', '1.0', '--choose-latent-p', '0.4', '--dh-depth', '2', '--print-every', '5000', '--topk', '30', '--use-negative-edge', '--scgpt-cache-path', 'scgpt_cache_norman.h5', '--result-path', './result_scgpt']' returned non-zero exit status 1.
57
+ ==========================================
58
+ Finished: Sun Mar 15 00:12:56 JST 2026
59
+ ==========================================
transfer/code/CCFM/logs/ccfm_1gpu_cached_5404526.out ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/200000 [00:00<?, ?it/s]Traceback (most recent call last):
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  0%| | 0/200000 [00:42<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==========================================
2
+ Job ID: 5404526
3
+ Job Name: ccfm_1gpu_cached
4
+ Start: Sun Mar 15 01:17:06 JST 2026
5
+ Node: b0027
6
+ GPU: NVIDIA H100, 95830 MiB
7
+ Run: CCFM 1GPU cached scGPT
8
+ ==========================================
9
+ The following values were not passed to `accelerate launch` and had defaults used instead:
10
+ `--num_machines` was set to a value of `1`
11
+ `--mixed_precision` was set to a value of `'no'`
12
+ `--dynamo_backend` was set to a value of `'no'`
13
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
14
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
15
+ warnings.warn("flash_attn is not installed")
16
+ WARNING:accelerate.utils.other:[RANK 0] Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
17
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
18
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
19
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
20
+ adata.uns["hvg"] = {"flavor": flavor}
21
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
22
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
23
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
24
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
25
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
26
+ warnings.warn(
27
+ CascadedFlowConfig(model_type='cascaded', batch_size=48, ntoken=512, d_model=128, nhead=8, nlayers=4, lr=5e-05, steps=200000, eta_min=1e-06, devices='1', test_only=False, data_name='norman', perturbation_function='crisper', noise_type='Gaussian', poisson_alpha=0.8, poisson_target_sum=-1, print_every=5000, mode='predict_y', result_path='./result_scgpt', fusion_method='differential_perceiver', infer_top_gene=1000, n_top_genes=5000, checkpoint_path='', gamma=0.5, split_method='additive', use_mmd_loss=True, fold=1, use_negative_edge=True, topk=30, scgpt_dim=512, bottleneck_dim=128, latent_weight=1.0, choose_latent_p=0.4, target_std=1.0, dh_depth=2, warmup_batches=200, scgpt_model_dir='transfer/data/scGPT_pretrained', scgpt_max_seq_len=1200, scgpt_cache_path='scgpt_cache_norman.h5', latent_steps=20, expr_steps=20)
28
+ Converted var_names to gene symbols, sample: ['RP11-34P13.8', 'RP11-54O7.3', 'SAMD11', 'PERM1', 'HES4']
29
+ Warning: ctrl is not in the gene names
30
+ ##### loading vocab from file #####
31
+ Loaded 159/163 pretrained parameters
32
+ Using pre-extracted scGPT cache: scgpt_cache_norman.h5
33
+ Cache shape: (91205, 5035, 512), cells: 91205
34
+
35
  0%| | 0/200000 [00:00<?, ?it/s]Traceback (most recent call last):
36
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 325, in <module>
37
+ cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
38
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
39
+ raw = self.features[row_indices] # (B, G_full, D) as numpy
40
+ ~~~~~~~~~~~~~^^^^^^^^^^^^^
41
+ File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
42
+ File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
43
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
44
+ selection = sel.select(self.shape, args, dataset=self)
45
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
46
+ return selector.make_selection(args)
47
+ ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
48
+ File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
49
+ File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
50
+ File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
51
+ File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
52
+ TypeError: Indexing elements must be in increasing order
53
+
54
  0%| | 0/200000 [00:42<?, ?it/s]
55
+ Traceback (most recent call last):
56
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/accelerate", line 8, in <module>
57
+ sys.exit(main())
58
+ ~~~~^^
59
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/accelerate_cli.py", line 50, in main
60
+ args.func(args)
61
+ ~~~~~~~~~^^^^^^
62
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1405, in launch_command
63
+ simple_launcher(args)
64
+ ~~~~~~~~~~~~~~~^^^^^^
65
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 993, in simple_launcher
66
+ raise subprocess.CalledProcessError(returncode=process.returncode, cmd=cmd)
67
+ subprocess.CalledProcessError: Command '['/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/python', 'scripts/run_cascaded.py', '--data-name', 'norman', '--d-model', '128', '--nhead', '8', '--nlayers', '4', '--batch-size', '48', '--lr', '5e-5', '--steps', '200000', '--fusion-method', 'differential_perceiver', '--perturbation-function', 'crisper', '--noise-type', 'Gaussian', '--infer-top-gene', '1000', '--n-top-genes', '5000', '--use-mmd-loss', '--gamma', '0.5', '--split-method', 'additive', '--fold', '1', '--scgpt-dim', '512', '--bottleneck-dim', '128', '--latent-weight', '1.0', '--choose-latent-p', '0.4', '--dh-depth', '2', '--print-every', '5000', '--topk', '30', '--use-negative-edge', '--scgpt-cache-path', 'scgpt_cache_norman.h5', '--result-path', './result_scgpt']' returned non-zero exit status 1.
68
+ ==========================================
69
+ Finished: Sun Mar 15 01:19:22 JST 2026
70
+ ==========================================
transfer/code/CCFM/logs/ccfm_1gpu_online_5404417.out ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf48151b252a3896e8ae3132cae1e9fafbd433de265dd15f653765d9919a5ea1
3
+ size 34640252
transfer/code/CCFM/logs/ccfm_topk30_neg_5402619.out ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==========================================
2
+ Job ID: 5402619
3
+ Job Name: ccfm_topk30_neg
4
+ Start: Sat Mar 14 14:34:35 JST 2026
5
+ Node: b0018
6
+ GPU: NVIDIA H100, 95830 MiB
7
+ NVIDIA H100, 95830 MiB
8
+ NVIDIA H100, 95830 MiB
9
+ NVIDIA H100, 95830 MiB
10
+ Run: CCFM (topk=30, use_negative_edge=True)
11
+ ==========================================
12
+ The following values were not passed to `accelerate launch` and had defaults used instead:
13
+ More than one GPU was found, enabling multi-GPU training.
14
+ If this was unintended please pass in `--num_processes=1`.
15
+ `--num_machines` was set to a value of `1`
16
+ `--mixed_precision` was set to a value of `'no'`
17
+ `--dynamo_backend` was set to a value of `'no'`
18
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
19
+ [W314 14:35:31.809931854 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
20
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
21
+ warnings.warn("flash_attn is not installed")
22
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
23
+ warnings.warn("flash_attn is not installed")
24
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
25
+ warnings.warn("flash_attn is not installed")
26
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
27
+ warnings.warn("flash_attn is not installed")
28
+ [W314 14:36:13.805735826 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
29
+ [W314 14:36:13.808311201 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
30
+ [W314 14:36:13.810035709 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
31
+ [W314 14:36:13.815100046 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
32
+ WARNING:accelerate.utils.other:[RANK 0] Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
33
+ CascadedFlowConfig(model_type='cascaded', batch_size=48, ntoken=512, d_model=128, nhead=8, nlayers=4, lr=5e-05, steps=200000, eta_min=1e-06, devices='1', test_only=False, data_name='norman', perturbation_function='crisper', noise_type='Gaussian', poisson_alpha=0.8, poisson_target_sum=-1, print_every=10000, mode='predict_y', result_path='./result', fusion_method='differential_perceiver', infer_top_gene=1000, n_top_genes=5000, checkpoint_path='', gamma=0.5, split_method='additive', use_mmd_loss=True, fold=1, use_negative_edge=True, topk=30, scgpt_dim=512, bottleneck_dim=128, latent_weight=1.0, choose_latent_p=0.4, target_std=1.0, dh_depth=2, warmup_batches=200, scgpt_model_dir='transfer/data/scGPT_pretrained', scgpt_max_seq_len=1200, scgpt_cache_path='', latent_steps=20, expr_steps=20)
34
+ [rank0]: Traceback (most recent call last):
35
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 146, in <module>
36
+ [rank0]: os.makedirs(save_path, exist_ok=True)
37
+ [rank0]: ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^
38
+ [rank0]: File "<frozen os>", line 227, in makedirs
39
+ [rank0]: OSError: [Errno 36] File name too long: './result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4-use_negative_edge_True-topk_30-scgpt_online'
40
+ [rank0]:[W314 14:36:18.955909336 ProcessGroupNCCL.cpp:1553] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
41
+ W0314 14:36:24.320000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 201 closing signal SIGTERM
42
+ W0314 14:36:24.322000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 202 closing signal SIGTERM
43
+ W0314 14:36:24.322000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 203 closing signal SIGTERM
44
+ E0314 14:36:24.687000 128 torch/distributed/elastic/multiprocessing/api.py:984] failed (exitcode: 1) local_rank: 0 (pid: 200) of binary: /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/python
45
+ Traceback (most recent call last):
46
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/accelerate", line 8, in <module>
47
+ sys.exit(main())
48
+ ~~~~^^
49
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/accelerate_cli.py", line 50, in main
50
+ args.func(args)
51
+ ~~~~~~~~~^^^^^^
52
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1396, in launch_command
53
+ multi_gpu_launcher(args)
54
+ ~~~~~~~~~~~~~~~~~~^^^^^^
55
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1023, in multi_gpu_launcher
56
+ distrib_run.run(args)
57
+ ~~~~~~~~~~~~~~~^^^^^^
58
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/run.py", line 982, in run
59
+ elastic_launch(
60
+ ~~~~~~~~~~~~~~~
61
+ config=config,
62
+ ~~~~~~~~~~~~~~
63
+ entrypoint=cmd,
64
+ ~~~~~~~~~~~~~~~
65
+ )(*cmd_args)
66
+ ~^^^^^^^^^^^
67
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/launcher/api.py", line 170, in __call__
68
+ return launch_agent(self._config, self._entrypoint, list(args))
69
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/launcher/api.py", line 317, in launch_agent
70
+ raise ChildFailedError(
71
+ ...<2 lines>...
72
+ )
73
+ torch.distributed.elastic.multiprocessing.errors.ChildFailedError:
74
+ ============================================================
75
+ scripts/run_cascaded.py FAILED
76
+ ------------------------------------------------------------
77
+ Failures:
78
+ [1]:
79
+ time : 2026-03-14_14:36:24
80
+ host : b0018
81
+ rank : 1 (local_rank: 1)
82
+ exitcode : -15 (pid: 201)
83
+ error_file: <N/A>
84
+ traceback : Signal 15 (SIGTERM) received by PID 201
85
+ [2]:
86
+ time : 2026-03-14_14:36:24
87
+ host : b0018
88
+ rank : 2 (local_rank: 2)
89
+ exitcode : -15 (pid: 202)
90
+ error_file: <N/A>
91
+ traceback : Signal 15 (SIGTERM) received by PID 202
92
+ [3]:
93
+ time : 2026-03-14_14:36:24
94
+ host : b0018
95
+ rank : 3 (local_rank: 3)
96
+ exitcode : -15 (pid: 203)
97
+ error_file: <N/A>
98
+ traceback : Signal 15 (SIGTERM) received by PID 203
99
+ ------------------------------------------------------------
100
+ Root Cause (first observed failure):
101
+ [0]:
102
+ time : 2026-03-14_14:36:24
103
+ host : b0018
104
+ rank : 0 (local_rank: 0)
105
+ exitcode : 1 (pid: 200)
106
+ error_file: <N/A>
107
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
108
+ ============================================================
109
+ ==========================================
110
+ Finished: Sat Mar 14 14:36:25 JST 2026
111
+ ==========================================
transfer/code/CCFM/logs/ccfm_v2_5404727.out ADDED
@@ -0,0 +1,214 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/200000 [00:00<?, ?it/s]
1
  0%| | 0/200000 [00:00<?, ?it/s]
2
  0%| | 0/200000 [00:00<?, ?it/s]
3
  0%| | 0/200000 [00:00<?, ?it/s]/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/transformer.py:531: UserWarning: The PyTorch API of nested tensors is in prototype stage and will change in the near future. We recommend specifying layout=torch.jagged when constructing a nested tensor, as this layout receives active development, has better operator coverage, and works with torch.compile. (Triggered internally at /pytorch/aten/src/ATen/NestedTensorImpl.cpp:178.)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  0%| | 0/200000 [00:53<?, ?it/s]
 
5
  0%| | 0/200000 [00:53<?, ?it/s]
 
6
  0%| | 0/200000 [00:53<?, ?it/s]
 
7
  0%| | 0/200000 [00:53<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==========================================
2
+ Job ID: 5404727
3
+ Job Name: ccfm_v2
4
+ Start: Sun Mar 15 04:49:31 JST 2026
5
+ Node: b0019
6
+ GPU: NVIDIA H100, 95830 MiB
7
+ NVIDIA H100, 95830 MiB
8
+ NVIDIA H100, 95830 MiB
9
+ NVIDIA H100, 95830 MiB
10
+ Run: CCFM v2 (KV fix + loss fix + EMA + RK4 + logit-normal + warmup)
11
+ ==========================================
12
+ The following values were not passed to `accelerate launch` and had defaults used instead:
13
+ More than one GPU was found, enabling multi-GPU training.
14
+ If this was unintended please pass in `--num_processes=1`.
15
+ `--num_machines` was set to a value of `1`
16
+ `--mixed_precision` was set to a value of `'no'`
17
+ `--dynamo_backend` was set to a value of `'no'`
18
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
19
+ [W315 04:50:08.654423221 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
20
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
21
+ warnings.warn("flash_attn is not installed")
22
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
23
+ warnings.warn("flash_attn is not installed")
24
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
25
+ warnings.warn("flash_attn is not installed")
26
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
27
+ warnings.warn("flash_attn is not installed")
28
+ [W315 04:50:33.963867213 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
29
+ [W315 04:50:33.963864886 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
30
+ [W315 04:50:33.963886032 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
31
+ [W315 04:50:33.963888124 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
32
+ WARNING:accelerate.utils.other:[RANK 0] Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
33
+ CascadedFlowConfig(model_type='cascaded', batch_size=48, ntoken=512, d_model=128, nhead=8, nlayers=4, lr=5e-05, steps=200000, eta_min=1e-06, devices='1', test_only=False, data_name='norman', perturbation_function='crisper', noise_type='Gaussian', poisson_alpha=0.8, poisson_target_sum=-1, print_every=10000, mode='predict_y', result_path='./result', fusion_method='differential_perceiver', infer_top_gene=1000, n_top_genes=5000, checkpoint_path='', gamma=0.5, split_method='additive', use_mmd_loss=True, fold=1, use_negative_edge=True, topk=30, scgpt_dim=512, bottleneck_dim=128, latent_weight=1.0, choose_latent_p=0.4, target_std=1.0, dh_depth=2, warmup_batches=200, ema_decay=0.9999, t_sample_mode='logit_normal', t_expr_mean=0.0, t_expr_std=1.0, t_latent_mean=0.0, t_latent_std=1.0, warmup_steps=2000, scgpt_model_dir='transfer/data/scGPT_pretrained', scgpt_max_seq_len=1200, scgpt_cache_path='', latent_steps=20, expr_steps=20, ode_method='rk4')
34
+ Converted var_names to gene symbols, sample: ['RP11-34P13.8', 'RP11-54O7.3', 'SAMD11', 'PERM1', 'HES4']
35
+ Warning: ctrl is not in the gene namesWarning: ctrl is not in the gene names
36
+
37
+ Warning: ctrl is not in the gene names
38
+ Warning: ctrl is not in the gene names
39
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
40
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
41
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
42
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
43
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
44
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
45
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
46
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
47
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
48
+ adata.uns["hvg"] = {"flavor": flavor}
49
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
50
+ adata.uns["hvg"] = {"flavor": flavor}
51
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
52
+ adata.uns["hvg"] = {"flavor": flavor}
53
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
54
+ adata.uns["hvg"] = {"flavor": flavor}
55
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
56
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
57
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
58
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
59
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
60
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
61
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
62
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
63
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
64
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
65
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
66
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
67
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
68
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
69
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
70
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
71
+ ##### loading vocab from file #####
72
+ ##### loading vocab from file #####
73
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
74
+ warnings.warn(
75
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
76
+ warnings.warn(
77
+ ##### loading vocab from file #####
78
+ ##### loading vocab from file #####
79
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
80
+ warnings.warn(
81
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
82
+ warnings.warn(
83
+ Loaded 159/163 pretrained parameters
84
+ Loaded 159/163 pretrained parameters
85
+ Loaded 159/163 pretrained parameters
86
+ Loaded 159/163 pretrained parameters
87
+
88
  0%| | 0/200000 [00:00<?, ?it/s]
89
  0%| | 0/200000 [00:00<?, ?it/s]
90
  0%| | 0/200000 [00:00<?, ?it/s]
91
  0%| | 0/200000 [00:00<?, ?it/s]/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/transformer.py:531: UserWarning: The PyTorch API of nested tensors is in prototype stage and will change in the near future. We recommend specifying layout=torch.jagged when constructing a nested tensor, as this layout receives active development, has better operator coverage, and works with torch.compile. (Triggered internally at /pytorch/aten/src/ATen/NestedTensorImpl.cpp:178.)
92
+ output = torch._nested_tensor_from_mask(
93
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/transformer.py:531: UserWarning: The PyTorch API of nested tensors is in prototype stage and will change in the near future. We recommend specifying layout=torch.jagged when constructing a nested tensor, as this layout receives active development, has better operator coverage, and works with torch.compile. (Triggered internally at /pytorch/aten/src/ATen/NestedTensorImpl.cpp:178.)
94
+ output = torch._nested_tensor_from_mask(
95
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/transformer.py:531: UserWarning: The PyTorch API of nested tensors is in prototype stage and will change in the near future. We recommend specifying layout=torch.jagged when constructing a nested tensor, as this layout receives active development, has better operator coverage, and works with torch.compile. (Triggered internally at /pytorch/aten/src/ATen/NestedTensorImpl.cpp:178.)
96
+ output = torch._nested_tensor_from_mask(
97
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/transformer.py:531: UserWarning: The PyTorch API of nested tensors is in prototype stage and will change in the near future. We recommend specifying layout=torch.jagged when constructing a nested tensor, as this layout receives active development, has better operator coverage, and works with torch.compile. (Triggered internally at /pytorch/aten/src/ATen/NestedTensorImpl.cpp:178.)
98
+ output = torch._nested_tensor_from_mask(
99
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/linear.py:134: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 48 k 256 mat1_ld 256 mat2_ld 256 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
100
+ return F.linear(input, self.weight, self.bias)
101
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/linear.py:134: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 48 k 256 mat1_ld 256 mat2_ld 256 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
102
+ return F.linear(input, self.weight, self.bias)
103
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/linear.py:134: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 48 k 256 mat1_ld 256 mat2_ld 256 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
104
+ return F.linear(input, self.weight, self.bias)
105
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/linear.py:134: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 48 k 256 mat1_ld 256 mat2_ld 256 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
106
+ return F.linear(input, self.weight, self.bias)
107
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/linear.py:134: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 48 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
108
+ return F.linear(input, self.weight, self.bias)
109
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/linear.py:134: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 48 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
110
+ return F.linear(input, self.weight, self.bias)
111
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/linear.py:134: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 48 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
112
+ return F.linear(input, self.weight, self.bias)
113
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/linear.py:134: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 48 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
114
+ return F.linear(input, self.weight, self.bias)
115
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/functional.py:6637: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 96 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
116
+ attn_output = linear(attn_output, out_proj_weight, out_proj_bias)
117
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/functional.py:6637: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 96 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
118
+ attn_output = linear(attn_output, out_proj_weight, out_proj_bias)
119
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/functional.py:6637: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 96 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
120
+ attn_output = linear(attn_output, out_proj_weight, out_proj_bias)
121
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/functional.py:6637: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 96 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
122
+ attn_output = linear(attn_output, out_proj_weight, out_proj_bias)
123
+ [rank2]: Traceback (most recent call last):
124
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 372, in <module>
125
+ [rank2]: ema_p.lerp_(model_p.data, 1 - decay)
126
+ [rank2]: ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^
127
+ [rank2]: RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:2 and cpu!
128
+ [rank1]: Traceback (most recent call last):
129
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 372, in <module>
130
+ [rank1]: ema_p.lerp_(model_p.data, 1 - decay)
131
+ [rank1]: ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^
132
+ [rank1]: RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:1 and cpu!
133
+ [rank0]: Traceback (most recent call last):
134
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 372, in <module>
135
+ [rank0]: ema_p.lerp_(model_p.data, 1 - decay)
136
+ [rank0]: ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^
137
+ [rank0]: RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cpu!
138
+ [rank3]: Traceback (most recent call last):
139
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 372, in <module>
140
+ [rank3]: ema_p.lerp_(model_p.data, 1 - decay)
141
+ [rank3]: ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^
142
+ [rank3]: RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:3 and cpu!
143
+ [rank0]:[W315 04:52:01.076850828 ProcessGroupNCCL.cpp:1553] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
144
+
145
  0%| | 0/200000 [00:53<?, ?it/s]
146
+
147
  0%| | 0/200000 [00:53<?, ?it/s]
148
+
149
  0%| | 0/200000 [00:53<?, ?it/s]
150
+
151
  0%| | 0/200000 [00:53<?, ?it/s]
152
+ W0315 04:52:07.763000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 200 closing signal SIGTERM
153
+ W0315 04:52:07.765000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 201 closing signal SIGTERM
154
+ W0315 04:52:07.765000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 202 closing signal SIGTERM
155
+ E0315 04:52:08.130000 128 torch/distributed/elastic/multiprocessing/api.py:984] failed (exitcode: 1) local_rank: 3 (pid: 203) of binary: /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/python
156
+ Traceback (most recent call last):
157
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/accelerate", line 8, in <module>
158
+ sys.exit(main())
159
+ ~~~~^^
160
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/accelerate_cli.py", line 50, in main
161
+ args.func(args)
162
+ ~~~~~~~~~^^^^^^
163
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1396, in launch_command
164
+ multi_gpu_launcher(args)
165
+ ~~~~~~~~~~~~~~~~~~^^^^^^
166
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1023, in multi_gpu_launcher
167
+ distrib_run.run(args)
168
+ ~~~~~~~~~~~~~~~^^^^^^
169
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/run.py", line 982, in run
170
+ elastic_launch(
171
+ ~~~~~~~~~~~~~~~
172
+ config=config,
173
+ ~~~~~~~~~~~~~~
174
+ entrypoint=cmd,
175
+ ~~~~~~~~~~~~~~~
176
+ )(*cmd_args)
177
+ ~^^^^^^^^^^^
178
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/launcher/api.py", line 170, in __call__
179
+ return launch_agent(self._config, self._entrypoint, list(args))
180
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/launcher/api.py", line 317, in launch_agent
181
+ raise ChildFailedError(
182
+ ...<2 lines>...
183
+ )
184
+ torch.distributed.elastic.multiprocessing.errors.ChildFailedError:
185
+ ============================================================
186
+ scripts/run_cascaded.py FAILED
187
+ ------------------------------------------------------------
188
+ Failures:
189
+ [1]:
190
+ time : 2026-03-15_04:52:08
191
+ host : b0019
192
+ rank : 0 (local_rank: 0)
193
+ exitcode : 1 (pid: 200)
194
+ error_file: <N/A>
195
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
196
+ [2]:
197
+ time : 2026-03-15_04:52:08
198
+ host : b0019
199
+ rank : 1 (local_rank: 1)
200
+ exitcode : 1 (pid: 201)
201
+ error_file: <N/A>
202
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
203
+ [3]:
204
+ time : 2026-03-15_04:52:08
205
+ host : b0019
206
+ rank : 2 (local_rank: 2)
207
+ exitcode : 1 (pid: 202)
208
+ error_file: <N/A>
209
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
210
+ ------------------------------------------------------------
211
+ Root Cause (first observed failure):
212
+ [0]:
213
+ time : 2026-03-15_04:52:07
214
+ host : b0019
215
+ rank : 3 (local_rank: 3)
216
+ exitcode : 1 (pid: 203)
217
+ error_file: <N/A>
218
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
219
+ ============================================================
220
+ ==========================================
221
+ Finished: Sun Mar 15 04:52:09 JST 2026
222
+ ==========================================
transfer/code/CCFM/logs/ccfm_v2_5404735.out ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78ae507c08bcf68993af70a80ee5dc04ea4c1b5236c845f4c2da3d8cb3356573
3
+ size 81641352
transfer/code/CCFM/logs/ccfm_v2_cached_5404728.out ADDED
@@ -0,0 +1,236 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/200000 [00:00<?, ?it/s]
1
  0%| | 0/200000 [00:00<?, ?it/s]
2
  0%| | 0/200000 [00:00<?, ?it/s]
3
  0%| | 0/200000 [00:00<?, ?it/s][rank3]: Traceback (most recent call last):
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  0%| | 0/200000 [00:42<?, ?it/s]
 
5
  0%| | 0/200000 [00:42<?, ?it/s]
 
6
  0%| | 0/200000 [00:42<?, ?it/s]
 
7
  0%| | 0/200000 [00:42<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==========================================
2
+ Job ID: 5404728
3
+ Job Name: ccfm_v2_cached
4
+ Start: Sun Mar 15 04:52:33 JST 2026
5
+ Node: b0021
6
+ GPU: NVIDIA H100, 95830 MiB
7
+ NVIDIA H100, 95830 MiB
8
+ NVIDIA H100, 95830 MiB
9
+ NVIDIA H100, 95830 MiB
10
+ Run: CCFM v2 cached (KV fix + loss fix + EMA + RK4 + logit-normal + warmup + scGPT cache)
11
+ ==========================================
12
+ The following values were not passed to `accelerate launch` and had defaults used instead:
13
+ More than one GPU was found, enabling multi-GPU training.
14
+ If this was unintended please pass in `--num_processes=1`.
15
+ `--num_machines` was set to a value of `1`
16
+ `--mixed_precision` was set to a value of `'no'`
17
+ `--dynamo_backend` was set to a value of `'no'`
18
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
19
+ [W315 04:52:57.426142473 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
20
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
21
+ warnings.warn("flash_attn is not installed")
22
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
23
+ warnings.warn("flash_attn is not installed")
24
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
25
+ warnings.warn("flash_attn is not installed")
26
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
27
+ warnings.warn("flash_attn is not installed")
28
+ [W315 04:53:16.185740786 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
29
+ [W315 04:53:16.441379224 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
30
+ [W315 04:53:16.447651346 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
31
+ [W315 04:53:16.490639280 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
32
+ WARNING:accelerate.utils.other:[RANK 0] Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
33
+ CascadedFlowConfig(model_type='cascaded', batch_size=48, ntoken=512, d_model=128, nhead=8, nlayers=4, lr=5e-05, steps=200000, eta_min=1e-06, devices='1', test_only=False, data_name='norman', perturbation_function='crisper', noise_type='Gaussian', poisson_alpha=0.8, poisson_target_sum=-1, print_every=10000, mode='predict_y', result_path='./result', fusion_method='differential_perceiver', infer_top_gene=1000, n_top_genes=5000, checkpoint_path='', gamma=0.5, split_method='additive', use_mmd_loss=True, fold=1, use_negative_edge=True, topk=30, scgpt_dim=512, bottleneck_dim=128, latent_weight=1.0, choose_latent_p=0.4, target_std=1.0, dh_depth=2, warmup_batches=200, ema_decay=0.9999, t_sample_mode='logit_normal', t_expr_mean=0.0, t_expr_std=1.0, t_latent_mean=0.0, t_latent_std=1.0, warmup_steps=2000, scgpt_model_dir='transfer/data/scGPT_pretrained', scgpt_max_seq_len=1200, scgpt_cache_path='scgpt_cache_norman.h5', latent_steps=20, expr_steps=20, ode_method='rk4')
34
+ Converted var_names to gene symbols, sample: ['RP11-34P13.8', 'RP11-54O7.3', 'SAMD11', 'PERM1', 'HES4']
35
+ Warning: ctrl is not in the gene namesWarning: ctrl is not in the gene names
36
+ Warning: ctrl is not in the gene names
37
+
38
+ Warning: ctrl is not in the gene names
39
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
40
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
41
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
42
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
43
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
44
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
45
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
46
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
47
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
48
+ adata.uns["hvg"] = {"flavor": flavor}
49
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
50
+ adata.uns["hvg"] = {"flavor": flavor}
51
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
52
+ adata.uns["hvg"] = {"flavor": flavor}
53
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
54
+ adata.uns["hvg"] = {"flavor": flavor}
55
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
56
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
57
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
58
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
59
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
60
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
61
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
62
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
63
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
64
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
65
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
66
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
67
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
68
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
69
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
70
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
71
+ ##### loading vocab from file #####
72
+ ##### loading vocab from file #####
73
+ ##### loading vocab from file #####
74
+ ##### loading vocab from file #####
75
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
76
+ warnings.warn(
77
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
78
+ warnings.warn(
79
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
80
+ warnings.warn(
81
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
82
+ warnings.warn(
83
+ Loaded 159/163 pretrained parameters
84
+ Loaded 159/163 pretrained parameters
85
+ Loaded 159/163 pretrained parameters
86
+ Loaded 159/163 pretrained parameters
87
+ Using pre-extracted scGPT cache: scgpt_cache_norman.h5
88
+ Cache shape: (91205, 5035, 512), cells: 91205
89
+
90
  0%| | 0/200000 [00:00<?, ?it/s]
91
  0%| | 0/200000 [00:00<?, ?it/s]
92
  0%| | 0/200000 [00:00<?, ?it/s]
93
  0%| | 0/200000 [00:00<?, ?it/s][rank3]: Traceback (most recent call last):
94
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
95
+ [rank3]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
96
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
97
+ [rank3]: raw = self.features[row_indices] # (B, G_full, D) as numpy
98
+ [rank3]: ~~~~~~~~~~~~~^^^^^^^^^^^^^
99
+ [rank3]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
100
+ [rank3]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
101
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
102
+ [rank3]: selection = sel.select(self.shape, args, dataset=self)
103
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
104
+ [rank3]: return selector.make_selection(args)
105
+ [rank3]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
106
+ [rank3]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
107
+ [rank3]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
108
+ [rank3]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
109
+ [rank3]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
110
+ [rank3]: TypeError: Indexing elements must be in increasing order
111
+ [rank0]: Traceback (most recent call last):
112
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
113
+ [rank0]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
114
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
115
+ [rank0]: raw = self.features[row_indices] # (B, G_full, D) as numpy
116
+ [rank0]: ~~~~~~~~~~~~~^^^^^^^^^^^^^
117
+ [rank0]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
118
+ [rank0]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
119
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
120
+ [rank0]: selection = sel.select(self.shape, args, dataset=self)
121
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
122
+ [rank0]: return selector.make_selection(args)
123
+ [rank0]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
124
+ [rank0]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
125
+ [rank0]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
126
+ [rank0]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
127
+ [rank0]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
128
+ [rank0]: TypeError: Indexing elements must be in increasing order
129
+ [rank1]: Traceback (most recent call last):
130
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
131
+ [rank1]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
132
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
133
+ [rank1]: raw = self.features[row_indices] # (B, G_full, D) as numpy
134
+ [rank1]: ~~~~~~~~~~~~~^^^^^^^^^^^^^
135
+ [rank1]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
136
+ [rank1]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
137
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
138
+ [rank1]: selection = sel.select(self.shape, args, dataset=self)
139
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
140
+ [rank1]: return selector.make_selection(args)
141
+ [rank1]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
142
+ [rank1]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
143
+ [rank1]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
144
+ [rank1]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
145
+ [rank1]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
146
+ [rank1]: TypeError: Indexing elements must be in increasing order
147
+ [rank2]: Traceback (most recent call last):
148
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
149
+ [rank2]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
150
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
151
+ [rank2]: raw = self.features[row_indices] # (B, G_full, D) as numpy
152
+ [rank2]: ~~~~~~~~~~~~~^^^^^^^^^^^^^
153
+ [rank2]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
154
+ [rank2]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
155
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
156
+ [rank2]: selection = sel.select(self.shape, args, dataset=self)
157
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
158
+ [rank2]: return selector.make_selection(args)
159
+ [rank2]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
160
+ [rank2]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
161
+ [rank2]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
162
+ [rank2]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
163
+ [rank2]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
164
+ [rank2]: TypeError: Indexing elements must be in increasing order
165
+ [rank0]:[W315 04:54:29.650127266 ProcessGroupNCCL.cpp:1553] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
166
+
167
  0%| | 0/200000 [00:42<?, ?it/s]
168
+
169
  0%| | 0/200000 [00:42<?, ?it/s]
170
+
171
  0%| | 0/200000 [00:42<?, ?it/s]
172
+
173
  0%| | 0/200000 [00:42<?, ?it/s]
174
+ W0315 04:54:34.559000 129 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 201 closing signal SIGTERM
175
+ W0315 04:54:34.560000 129 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 202 closing signal SIGTERM
176
+ W0315 04:54:34.561000 129 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 204 closing signal SIGTERM
177
+ E0315 04:54:34.775000 129 torch/distributed/elastic/multiprocessing/api.py:984] failed (exitcode: 1) local_rank: 2 (pid: 203) of binary: /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/python
178
+ Traceback (most recent call last):
179
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/accelerate", line 8, in <module>
180
+ sys.exit(main())
181
+ ~~~~^^
182
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/accelerate_cli.py", line 50, in main
183
+ args.func(args)
184
+ ~~~~~~~~~^^^^^^
185
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1396, in launch_command
186
+ multi_gpu_launcher(args)
187
+ ~~~~~~~~~~~~~~~~~~^^^^^^
188
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1023, in multi_gpu_launcher
189
+ distrib_run.run(args)
190
+ ~~~~~~~~~~~~~~~^^^^^^
191
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/run.py", line 982, in run
192
+ elastic_launch(
193
+ ~~~~~~~~~~~~~~~
194
+ config=config,
195
+ ~~~~~~~~~~~~~~
196
+ entrypoint=cmd,
197
+ ~~~~~~~~~~~~~~~
198
+ )(*cmd_args)
199
+ ~^^^^^^^^^^^
200
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/launcher/api.py", line 170, in __call__
201
+ return launch_agent(self._config, self._entrypoint, list(args))
202
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/launcher/api.py", line 317, in launch_agent
203
+ raise ChildFailedError(
204
+ ...<2 lines>...
205
+ )
206
+ torch.distributed.elastic.multiprocessing.errors.ChildFailedError:
207
+ ============================================================
208
+ scripts/run_cascaded.py FAILED
209
+ ------------------------------------------------------------
210
+ Failures:
211
+ [1]:
212
+ time : 2026-03-15_04:54:34
213
+ host : b0021
214
+ rank : 0 (local_rank: 0)
215
+ exitcode : 1 (pid: 201)
216
+ error_file: <N/A>
217
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
218
+ [2]:
219
+ time : 2026-03-15_04:54:34
220
+ host : b0021
221
+ rank : 1 (local_rank: 1)
222
+ exitcode : 1 (pid: 202)
223
+ error_file: <N/A>
224
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
225
+ [3]:
226
+ time : 2026-03-15_04:54:34
227
+ host : b0021
228
+ rank : 3 (local_rank: 3)
229
+ exitcode : 1 (pid: 204)
230
+ error_file: <N/A>
231
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
232
+ ------------------------------------------------------------
233
+ Root Cause (first observed failure):
234
+ [0]:
235
+ time : 2026-03-15_04:54:34
236
+ host : b0021
237
+ rank : 2 (local_rank: 2)
238
+ exitcode : 1 (pid: 203)
239
+ error_file: <N/A>
240
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
241
+ ============================================================
242
+ ==========================================
243
+ Finished: Sun Mar 15 04:54:35 JST 2026
244
+ ==========================================
transfer/code/CCFM/logs/ccfm_v2_cached_5404736.out ADDED
@@ -0,0 +1,236 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/200000 [00:00<?, ?it/s]
1
  0%| | 0/200000 [00:00<?, ?it/s]
2
  0%| | 0/200000 [00:00<?, ?it/s]
3
  0%| | 0/200000 [00:00<?, ?it/s][rank0]: Traceback (most recent call last):
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  0%| | 0/200000 [00:42<?, ?it/s]
 
5
  0%| | 0/200000 [00:42<?, ?it/s]
 
6
  0%| | 0/200000 [00:42<?, ?it/s]
 
7
  0%| | 0/200000 [00:43<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==========================================
2
+ Job ID: 5404736
3
+ Job Name: ccfm_v2_cached
4
+ Start: Sun Mar 15 04:58:13 JST 2026
5
+ Node: b0021
6
+ GPU: NVIDIA H100, 95830 MiB
7
+ NVIDIA H100, 95830 MiB
8
+ NVIDIA H100, 95830 MiB
9
+ NVIDIA H100, 95830 MiB
10
+ Run: CCFM v2 cached (KV fix + loss fix + EMA + RK4 + logit-normal + warmup + scGPT cache)
11
+ ==========================================
12
+ The following values were not passed to `accelerate launch` and had defaults used instead:
13
+ More than one GPU was found, enabling multi-GPU training.
14
+ If this was unintended please pass in `--num_processes=1`.
15
+ `--num_machines` was set to a value of `1`
16
+ `--mixed_precision` was set to a value of `'no'`
17
+ `--dynamo_backend` was set to a value of `'no'`
18
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
19
+ [W315 04:58:44.021644370 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
20
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
21
+ warnings.warn("flash_attn is not installed")
22
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
23
+ warnings.warn("flash_attn is not installed")
24
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
25
+ warnings.warn("flash_attn is not installed")
26
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
27
+ warnings.warn("flash_attn is not installed")
28
+ [W315 04:59:02.163293234 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
29
+ [W315 04:59:02.432132628 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
30
+ [W315 04:59:02.432621699 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
31
+ [W315 04:59:02.442458875 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
32
+ WARNING:accelerate.utils.other:[RANK 0] Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
33
+ CascadedFlowConfig(model_type='cascaded', batch_size=48, ntoken=512, d_model=128, nhead=8, nlayers=4, lr=5e-05, steps=200000, eta_min=1e-06, devices='1', test_only=False, data_name='norman', perturbation_function='crisper', noise_type='Gaussian', poisson_alpha=0.8, poisson_target_sum=-1, print_every=10000, mode='predict_y', result_path='./result', fusion_method='differential_perceiver', infer_top_gene=1000, n_top_genes=5000, checkpoint_path='', gamma=0.5, split_method='additive', use_mmd_loss=True, fold=1, use_negative_edge=True, topk=30, scgpt_dim=512, bottleneck_dim=128, latent_weight=1.0, choose_latent_p=0.4, target_std=1.0, dh_depth=2, warmup_batches=200, ema_decay=0.9999, t_sample_mode='logit_normal', t_expr_mean=0.0, t_expr_std=1.0, t_latent_mean=0.0, t_latent_std=1.0, warmup_steps=2000, scgpt_model_dir='transfer/data/scGPT_pretrained', scgpt_max_seq_len=1200, scgpt_cache_path='scgpt_cache_norman.h5', latent_steps=20, expr_steps=20, ode_method='rk4')
34
+ Converted var_names to gene symbols, sample: ['RP11-34P13.8', 'RP11-54O7.3', 'SAMD11', 'PERM1', 'HES4']
35
+ Warning: ctrl is not in the gene names
36
+ Warning: ctrl is not in the gene namesWarning: ctrl is not in the gene names
37
+
38
+ Warning: ctrl is not in the gene names
39
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
40
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
41
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
42
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
43
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
44
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
45
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
46
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
47
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
48
+ adata.uns["hvg"] = {"flavor": flavor}
49
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
50
+ adata.uns["hvg"] = {"flavor": flavor}
51
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
52
+ adata.uns["hvg"] = {"flavor": flavor}
53
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
54
+ adata.uns["hvg"] = {"flavor": flavor}
55
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
56
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
57
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
58
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
59
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
60
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
61
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
62
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
63
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
64
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
65
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
66
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
67
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
68
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
69
+ ##### loading vocab from file #####
70
+ ##### loading vocab from file #####
71
+ ##### loading vocab from file #####
72
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
73
+ warnings.warn(
74
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
75
+ warnings.warn(
76
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
77
+ warnings.warn(
78
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
79
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
80
+ ##### loading vocab from file #####
81
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
82
+ warnings.warn(
83
+ Loaded 159/163 pretrained parameters
84
+ Loaded 159/163 pretrained parameters
85
+ Loaded 159/163 pretrained parameters
86
+ Using pre-extracted scGPT cache: scgpt_cache_norman.h5
87
+ Cache shape: (91205, 5035, 512), cells: 91205
88
+ Loaded 159/163 pretrained parameters
89
+
90
  0%| | 0/200000 [00:00<?, ?it/s]
91
  0%| | 0/200000 [00:00<?, ?it/s]
92
  0%| | 0/200000 [00:00<?, ?it/s]
93
  0%| | 0/200000 [00:00<?, ?it/s][rank0]: Traceback (most recent call last):
94
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
95
+ [rank0]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
96
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
97
+ [rank0]: raw = self.features[row_indices] # (B, G_full, D) as numpy
98
+ [rank0]: ~~~~~~~~~~~~~^^^^^^^^^^^^^
99
+ [rank0]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
100
+ [rank0]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
101
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
102
+ [rank0]: selection = sel.select(self.shape, args, dataset=self)
103
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
104
+ [rank0]: return selector.make_selection(args)
105
+ [rank0]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
106
+ [rank0]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
107
+ [rank0]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
108
+ [rank0]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
109
+ [rank0]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
110
+ [rank0]: TypeError: Indexing elements must be in increasing order
111
+ [rank2]: Traceback (most recent call last):
112
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
113
+ [rank2]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
114
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
115
+ [rank2]: raw = self.features[row_indices] # (B, G_full, D) as numpy
116
+ [rank2]: ~~~~~~~~~~~~~^^^^^^^^^^^^^
117
+ [rank2]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
118
+ [rank2]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
119
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
120
+ [rank2]: selection = sel.select(self.shape, args, dataset=self)
121
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
122
+ [rank2]: return selector.make_selection(args)
123
+ [rank2]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
124
+ [rank2]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
125
+ [rank2]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
126
+ [rank2]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
127
+ [rank2]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
128
+ [rank2]: TypeError: Indexing elements must be in increasing order
129
+ [rank1]: Traceback (most recent call last):
130
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
131
+ [rank1]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
132
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
133
+ [rank1]: raw = self.features[row_indices] # (B, G_full, D) as numpy
134
+ [rank1]: ~~~~~~~~~~~~~^^^^^^^^^^^^^
135
+ [rank1]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
136
+ [rank1]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
137
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
138
+ [rank1]: selection = sel.select(self.shape, args, dataset=self)
139
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
140
+ [rank1]: return selector.make_selection(args)
141
+ [rank1]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
142
+ [rank1]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
143
+ [rank1]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
144
+ [rank1]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
145
+ [rank1]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
146
+ [rank1]: TypeError: Indexing elements must be in increasing order
147
+ [rank3]: Traceback (most recent call last):
148
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
149
+ [rank3]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
150
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
151
+ [rank3]: raw = self.features[row_indices] # (B, G_full, D) as numpy
152
+ [rank3]: ~~~~~~~~~~~~~^^^^^^^^^^^^^
153
+ [rank3]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
154
+ [rank3]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
155
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
156
+ [rank3]: selection = sel.select(self.shape, args, dataset=self)
157
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
158
+ [rank3]: return selector.make_selection(args)
159
+ [rank3]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
160
+ [rank3]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
161
+ [rank3]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
162
+ [rank3]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
163
+ [rank3]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
164
+ [rank3]: TypeError: Indexing elements must be in increasing order
165
+ [rank0]:[W315 05:00:15.940375045 ProcessGroupNCCL.cpp:1553] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
166
+
167
  0%| | 0/200000 [00:42<?, ?it/s]
168
+
169
  0%| | 0/200000 [00:42<?, ?it/s]
170
+
171
  0%| | 0/200000 [00:42<?, ?it/s]
172
+
173
  0%| | 0/200000 [00:43<?, ?it/s]
174
+ W0315 05:00:19.854000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 200 closing signal SIGTERM
175
+ W0315 05:00:19.856000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 201 closing signal SIGTERM
176
+ W0315 05:00:19.856000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 203 closing signal SIGTERM
177
+ E0315 05:00:20.071000 128 torch/distributed/elastic/multiprocessing/api.py:984] failed (exitcode: 1) local_rank: 2 (pid: 202) of binary: /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/python
178
+ Traceback (most recent call last):
179
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/accelerate", line 8, in <module>
180
+ sys.exit(main())
181
+ ~~~~^^
182
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/accelerate_cli.py", line 50, in main
183
+ args.func(args)
184
+ ~~~~~~~~~^^^^^^
185
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1396, in launch_command
186
+ multi_gpu_launcher(args)
187
+ ~~~~~~~~~~~~~~~~~~^^^^^^
188
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1023, in multi_gpu_launcher
189
+ distrib_run.run(args)
190
+ ~~~~~~~~~~~~~~~^^^^^^
191
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/run.py", line 982, in run
192
+ elastic_launch(
193
+ ~~~~~~~~~~~~~~~
194
+ config=config,
195
+ ~~~~~~~~~~~~~~
196
+ entrypoint=cmd,
197
+ ~~~~~~~~~~~~~~~
198
+ )(*cmd_args)
199
+ ~^^^^^^^^^^^
200
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/launcher/api.py", line 170, in __call__
201
+ return launch_agent(self._config, self._entrypoint, list(args))
202
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/launcher/api.py", line 317, in launch_agent
203
+ raise ChildFailedError(
204
+ ...<2 lines>...
205
+ )
206
+ torch.distributed.elastic.multiprocessing.errors.ChildFailedError:
207
+ ============================================================
208
+ scripts/run_cascaded.py FAILED
209
+ ------------------------------------------------------------
210
+ Failures:
211
+ [1]:
212
+ time : 2026-03-15_05:00:20
213
+ host : b0021
214
+ rank : 0 (local_rank: 0)
215
+ exitcode : 1 (pid: 200)
216
+ error_file: <N/A>
217
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
218
+ [2]:
219
+ time : 2026-03-15_05:00:20
220
+ host : b0021
221
+ rank : 1 (local_rank: 1)
222
+ exitcode : 1 (pid: 201)
223
+ error_file: <N/A>
224
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
225
+ [3]:
226
+ time : 2026-03-15_05:00:20
227
+ host : b0021
228
+ rank : 3 (local_rank: 3)
229
+ exitcode : 1 (pid: 203)
230
+ error_file: <N/A>
231
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
232
+ ------------------------------------------------------------
233
+ Root Cause (first observed failure):
234
+ [0]:
235
+ time : 2026-03-15_05:00:19
236
+ host : b0021
237
+ rank : 2 (local_rank: 2)
238
+ exitcode : 1 (pid: 202)
239
+ error_file: <N/A>
240
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
241
+ ============================================================
242
+ ==========================================
243
+ Finished: Sun Mar 15 05:00:21 JST 2026
244
+ ==========================================
transfer/code/CCFM/logs/ccfm_v2_cached_5404737.out ADDED
@@ -0,0 +1,236 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/200000 [00:00<?, ?it/s]
1
  0%| | 0/200000 [00:00<?, ?it/s]
2
  0%| | 0/200000 [00:00<?, ?it/s]
3
  0%| | 0/200000 [00:00<?, ?it/s][rank0]: Traceback (most recent call last):
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  0%| | 0/200000 [00:42<?, ?it/s]
 
5
  0%| | 0/200000 [00:42<?, ?it/s]
 
6
  0%| | 0/200000 [00:42<?, ?it/s]
 
7
  0%| | 0/200000 [00:42<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==========================================
2
+ Job ID: 5404737
3
+ Job Name: ccfm_v2_cached
4
+ Start: Sun Mar 15 05:02:01 JST 2026
5
+ Node: b0021
6
+ GPU: NVIDIA H100, 95830 MiB
7
+ NVIDIA H100, 95830 MiB
8
+ NVIDIA H100, 95830 MiB
9
+ NVIDIA H100, 95830 MiB
10
+ Run: CCFM v2 cached (KV fix + loss fix + EMA + RK4 + logit-normal + warmup + scGPT cache)
11
+ ==========================================
12
+ The following values were not passed to `accelerate launch` and had defaults used instead:
13
+ More than one GPU was found, enabling multi-GPU training.
14
+ If this was unintended please pass in `--num_processes=1`.
15
+ `--num_machines` was set to a value of `1`
16
+ `--mixed_precision` was set to a value of `'no'`
17
+ `--dynamo_backend` was set to a value of `'no'`
18
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
19
+ [W315 05:02:27.190711515 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
20
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
21
+ warnings.warn("flash_attn is not installed")
22
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
23
+ warnings.warn("flash_attn is not installed")
24
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
25
+ warnings.warn("flash_attn is not installed")
26
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
27
+ warnings.warn("flash_attn is not installed")
28
+ [W315 05:02:46.969095263 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
29
+ [W315 05:02:46.996806598 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
30
+ [W315 05:02:46.020582775 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
31
+ [W315 05:02:46.024342649 socket.cpp:764] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
32
+ WARNING:accelerate.utils.other:[RANK 0] Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
33
+ CascadedFlowConfig(model_type='cascaded', batch_size=48, ntoken=512, d_model=128, nhead=8, nlayers=4, lr=5e-05, steps=200000, eta_min=1e-06, devices='1', test_only=False, data_name='norman', perturbation_function='crisper', noise_type='Gaussian', poisson_alpha=0.8, poisson_target_sum=-1, print_every=10000, mode='predict_y', result_path='./result', fusion_method='differential_perceiver', infer_top_gene=1000, n_top_genes=5000, checkpoint_path='', gamma=0.5, split_method='additive', use_mmd_loss=True, fold=1, use_negative_edge=True, topk=30, scgpt_dim=512, bottleneck_dim=128, latent_weight=1.0, choose_latent_p=0.4, target_std=1.0, dh_depth=2, warmup_batches=200, ema_decay=0.9999, t_sample_mode='logit_normal', t_expr_mean=0.0, t_expr_std=1.0, t_latent_mean=0.0, t_latent_std=1.0, warmup_steps=2000, scgpt_model_dir='transfer/data/scGPT_pretrained', scgpt_max_seq_len=1200, scgpt_cache_path='scgpt_cache_norman.h5', latent_steps=20, expr_steps=20, ode_method='rk4')
34
+ Converted var_names to gene symbols, sample: ['RP11-34P13.8', 'RP11-54O7.3', 'SAMD11', 'PERM1', 'HES4']
35
+ Warning: ctrl is not in the gene namesWarning: ctrl is not in the gene namesWarning: ctrl is not in the gene names
36
+
37
+
38
+ Warning: ctrl is not in the gene names
39
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
40
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
41
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
42
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
43
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
44
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
45
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
46
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
47
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
48
+ adata.uns["hvg"] = {"flavor": flavor}
49
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
50
+ adata.uns["hvg"] = {"flavor": flavor}
51
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
52
+ adata.uns["hvg"] = {"flavor": flavor}
53
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
54
+ adata.uns["hvg"] = {"flavor": flavor}
55
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
56
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
57
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
58
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
59
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
60
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
61
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
62
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
63
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
64
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
65
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
66
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
67
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
68
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
69
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
70
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
71
+ ##### loading vocab from file #####
72
+ ##### loading vocab from file #####
73
+ ##### loading vocab from file #####
74
+ ##### loading vocab from file #####
75
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
76
+ warnings.warn(
77
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
78
+ warnings.warn(
79
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
80
+ warnings.warn(
81
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
82
+ warnings.warn(
83
+ Loaded 159/163 pretrained parameters
84
+ Loaded 159/163 pretrained parameters
85
+ Loaded 159/163 pretrained parameters
86
+ Loaded 159/163 pretrained parameters
87
+ Using pre-extracted scGPT cache: scgpt_cache_norman.h5
88
+ Cache shape: (91205, 5035, 512), cells: 91205
89
+
90
  0%| | 0/200000 [00:00<?, ?it/s]
91
  0%| | 0/200000 [00:00<?, ?it/s]
92
  0%| | 0/200000 [00:00<?, ?it/s]
93
  0%| | 0/200000 [00:00<?, ?it/s][rank0]: Traceback (most recent call last):
94
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
95
+ [rank0]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
96
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
97
+ [rank0]: raw = self.features[sorted_indices.tolist()] # (B, G_full, D) as numpy
98
+ [rank0]: ~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^
99
+ [rank0]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
100
+ [rank0]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
101
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
102
+ [rank0]: selection = sel.select(self.shape, args, dataset=self)
103
+ [rank0]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
104
+ [rank0]: return selector.make_selection(args)
105
+ [rank0]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
106
+ [rank0]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
107
+ [rank0]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
108
+ [rank0]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
109
+ [rank0]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
110
+ [rank0]: TypeError: Indexing elements must be in increasing order
111
+ [rank1]: Traceback (most recent call last):
112
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
113
+ [rank1]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
114
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
115
+ [rank1]: raw = self.features[sorted_indices.tolist()] # (B, G_full, D) as numpy
116
+ [rank1]: ~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^
117
+ [rank1]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
118
+ [rank1]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
119
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
120
+ [rank1]: selection = sel.select(self.shape, args, dataset=self)
121
+ [rank1]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
122
+ [rank1]: return selector.make_selection(args)
123
+ [rank1]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
124
+ [rank1]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
125
+ [rank1]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
126
+ [rank1]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
127
+ [rank1]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
128
+ [rank1]: TypeError: Indexing elements must be in increasing order
129
+ [rank2]: Traceback (most recent call last):
130
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
131
+ [rank2]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
132
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
133
+ [rank2]: raw = self.features[sorted_indices.tolist()] # (B, G_full, D) as numpy
134
+ [rank2]: ~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^
135
+ [rank2]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
136
+ [rank2]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
137
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
138
+ [rank2]: selection = sel.select(self.shape, args, dataset=self)
139
+ [rank2]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
140
+ [rank2]: return selector.make_selection(args)
141
+ [rank2]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
142
+ [rank2]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
143
+ [rank2]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
144
+ [rank2]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
145
+ [rank2]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
146
+ [rank2]: TypeError: Indexing elements must be in increasing order
147
+ [rank3]: Traceback (most recent call last):
148
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scripts/run_cascaded.py", line 350, in <module>
149
+ [rank3]: cached_z_target = scgpt_cache.lookup(tgt_cell_names, input_gene_ids, device=device)
150
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_cache.py", line 55, in lookup
151
+ [rank3]: raw = self.features[sorted_indices.tolist()] # (B, G_full, D) as numpy
152
+ [rank3]: ~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^
153
+ [rank3]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
154
+ [rank3]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
155
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/dataset.py", line 945, in __getitem__
156
+ [rank3]: selection = sel.select(self.shape, args, dataset=self)
157
+ [rank3]: File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/h5py/_hl/selections.py", line 85, in select
158
+ [rank3]: return selector.make_selection(args)
159
+ [rank3]: ~~~~~~~~~~~~~~~~~~~~~~~^^^^^^
160
+ [rank3]: File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
161
+ [rank3]: File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
162
+ [rank3]: File "h5py/_selector.pyx", line 297, in h5py._selector.Selector.make_selection
163
+ [rank3]: File "h5py/_selector.pyx", line 216, in h5py._selector.Selector.apply_args
164
+ [rank3]: TypeError: Indexing elements must be in increasing order
165
+ [rank0]:[W315 05:04:02.931689961 ProcessGroupNCCL.cpp:1553] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
166
+
167
  0%| | 0/200000 [00:42<?, ?it/s]
168
+
169
  0%| | 0/200000 [00:42<?, ?it/s]
170
+
171
  0%| | 0/200000 [00:42<?, ?it/s]
172
+
173
  0%| | 0/200000 [00:42<?, ?it/s]
174
+ W0315 05:04:06.726000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 200 closing signal SIGTERM
175
+ W0315 05:04:06.728000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 201 closing signal SIGTERM
176
+ W0315 05:04:06.729000 128 torch/distributed/elastic/multiprocessing/api.py:1010] Sending process 203 closing signal SIGTERM
177
+ E0315 05:04:06.943000 128 torch/distributed/elastic/multiprocessing/api.py:984] failed (exitcode: 1) local_rank: 2 (pid: 202) of binary: /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/python
178
+ Traceback (most recent call last):
179
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/bin/accelerate", line 8, in <module>
180
+ sys.exit(main())
181
+ ~~~~^^
182
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/accelerate_cli.py", line 50, in main
183
+ args.func(args)
184
+ ~~~~~~~~~^^^^^^
185
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1396, in launch_command
186
+ multi_gpu_launcher(args)
187
+ ~~~~~~~~~~~~~~~~~~^^^^^^
188
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/accelerate/commands/launch.py", line 1023, in multi_gpu_launcher
189
+ distrib_run.run(args)
190
+ ~~~~~~~~~~~~~~~^^^^^^
191
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/run.py", line 982, in run
192
+ elastic_launch(
193
+ ~~~~~~~~~~~~~~~
194
+ config=config,
195
+ ~~~~~~~~~~~~~~
196
+ entrypoint=cmd,
197
+ ~~~~~~~~~~~~~~~
198
+ )(*cmd_args)
199
+ ~^^^^^^^^^^^
200
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/launcher/api.py", line 170, in __call__
201
+ return launch_agent(self._config, self._entrypoint, list(args))
202
+ File "/home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/distributed/launcher/api.py", line 317, in launch_agent
203
+ raise ChildFailedError(
204
+ ...<2 lines>...
205
+ )
206
+ torch.distributed.elastic.multiprocessing.errors.ChildFailedError:
207
+ ============================================================
208
+ scripts/run_cascaded.py FAILED
209
+ ------------------------------------------------------------
210
+ Failures:
211
+ [1]:
212
+ time : 2026-03-15_05:04:06
213
+ host : b0021
214
+ rank : 0 (local_rank: 0)
215
+ exitcode : 1 (pid: 200)
216
+ error_file: <N/A>
217
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
218
+ [2]:
219
+ time : 2026-03-15_05:04:06
220
+ host : b0021
221
+ rank : 1 (local_rank: 1)
222
+ exitcode : 1 (pid: 201)
223
+ error_file: <N/A>
224
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
225
+ [3]:
226
+ time : 2026-03-15_05:04:06
227
+ host : b0021
228
+ rank : 3 (local_rank: 3)
229
+ exitcode : 1 (pid: 203)
230
+ error_file: <N/A>
231
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
232
+ ------------------------------------------------------------
233
+ Root Cause (first observed failure):
234
+ [0]:
235
+ time : 2026-03-15_05:04:06
236
+ host : b0021
237
+ rank : 2 (local_rank: 2)
238
+ exitcode : 1 (pid: 202)
239
+ error_file: <N/A>
240
+ traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
241
+ ============================================================
242
+ ==========================================
243
+ Finished: Sun Mar 15 05:04:07 JST 2026
244
+ ==========================================
transfer/code/CCFM/logs/ccfm_v2_cached_5405774.out ADDED
The diff for this file is too large to render. See raw diff
 
transfer/code/CCFM/logs/ccfm_v2_resume_5406605.out ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22687c5738583ccac9adb3790fa51dd831d4e439a1254a6f63da303c2d3f52f3
3
+ size 119504495
transfer/code/CCFM/logs/eval_ccfm_v2_5406214.out ADDED
@@ -0,0 +1,437 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==========================================
2
+ Job ID: 5406214
3
+ Job Name: eval_ccfm_v2
4
+ Start: Mon Mar 16 00:48:18 JST 2026
5
+ Node: b0031
6
+ GPU: NVIDIA H100, 95830 MiB
7
+ Eval: CCFM v2 (online) iteration_100000
8
+ ==========================================
9
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
10
+ warnings.warn("flash_attn is not installed")
11
+ WARNING:accelerate.utils.other:[RANK 0] Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
12
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
13
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
14
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
15
+ adata.uns["hvg"] = {"flavor": flavor}
16
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:277: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
17
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
18
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:329: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
19
+ self.adata.obs['perturbation_covariates'] = self.adata.obs[perturbation_covariates].apply(lambda x: '+'.join(x), axis=1)
20
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
21
+ warnings.warn(
22
+ CascadedFlowConfig(model_type='cascaded', batch_size=48, ntoken=512, d_model=128, nhead=8, nlayers=4, lr=5e-05, steps=200000, eta_min=1e-06, devices='1', test_only=True, data_name='norman', perturbation_function='crisper', noise_type='Gaussian', poisson_alpha=0.8, poisson_target_sum=-1, print_every=10000, mode='predict_y', result_path='./result', fusion_method='differential_perceiver', infer_top_gene=1000, n_top_genes=5000, checkpoint_path='/home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_100000/checkpoint.pt', gamma=0.5, split_method='additive', use_mmd_loss=True, fold=1, use_negative_edge=True, topk=30, scgpt_dim=512, bottleneck_dim=128, latent_weight=1.0, choose_latent_p=0.4, target_std=1.0, dh_depth=2, warmup_batches=200, ema_decay=0.9999, t_sample_mode='logit_normal', t_expr_mean=0.0, t_expr_std=1.0, t_latent_mean=0.0, t_latent_std=1.0, warmup_steps=2000, scgpt_model_dir='transfer/data/scGPT_pretrained', scgpt_max_seq_len=1200, scgpt_cache_path='', latent_steps=20, expr_steps=20, ode_method='rk4')
23
+ Converted var_names to gene symbols, sample: ['RP11-34P13.8', 'RP11-54O7.3', 'SAMD11', 'PERM1', 'HES4']
24
+ Warning: ctrl is not in the gene names
25
+ ##### loading vocab from file #####
26
+ Loaded 159/163 pretrained parameters
27
+ loading /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_100000/checkpoint.pt checkpoint, iteration: 100000, eval_score: None
28
+ Test-only mode. Saving results to ./result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/eval_only
29
+ perturbation_name_list: 39
30
+
31
+ return F.linear(input, self.weight, self.bias)
32
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/linear.py:134: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 48 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
33
+ return F.linear(input, self.weight, self.bias)
34
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/functional.py:6637: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 96 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
35
+ attn_output = linear(attn_output, out_proj_weight, out_proj_bias)
36
+
37
+ return F.linear(input, self.weight, self.bias)
38
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/modules/linear.py:134: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 32 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
39
+ return F.linear(input, self.weight, self.bias)
40
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/torch/nn/functional.py:6637: UserWarning: gemm_and_bias error: CUBLAS_STATUS_NOT_INITIALIZED when calling cublasLtMatmul with transpose_mat1 1 transpose_mat2 0 m 128 n 64 k 128 mat1_ld 128 mat2_ld 128 result_ld 128 abType 0 cType 0 computeType 68 scaleType 0. Will attempt to recover by calling unfused cublas path. (Triggered internally at /pytorch/aten/src/ATen/cuda/CUDABlas.cpp:1765.)
41
+ attn_output = linear(attn_output, out_proj_weight, out_proj_bias)
42
+
43
+
44
+
45
+
46
+
47
+
48
+
49
+
50
+
51
+
52
+
53
+
54
+
55
+
56
+
57
+
58
+
59
+
60
+
61
+
62
+
63
+
64
+
65
+
66
+
67
+
68
+
69
+
70
+
71
+
72
+
73
+
74
+
75
+
76
+
77
+
78
+
79
+
80
+
81
+ /home/hp250092/ku50001222/miniconda3/lib/python3.13/functools.py:934: ImplicitModificationWarning: Transforming to str index.
82
+ return dispatch(args[0].__class__)(*args, **kw)
83
+ /home/hp250092/ku50001222/miniconda3/lib/python3.13/functools.py:934: ImplicitModificationWarning: Transforming to str index.
84
+ return dispatch(args[0].__class__)(*args, **kw)
85
+ WARNING:cell_eval._evaluator:Output directory ./cell-eval-outdir already exists, potential overwrite occurring
86
+ INFO:cell_eval.utils:Data appears to be log1p normalized (decimals detected, range [0.00, 8.82])
87
+ INFO:cell_eval._evaluator:Input is found to be log-normalized already - skipping transformation.
88
+ INFO:cell_eval.utils:Data appears to be log1p normalized (decimals detected, range [0.00, 9.75])
89
+ INFO:cell_eval._evaluator:Input is found to be log-normalized already - skipping transformation.
90
+ INFO:cell_eval._evaluator:Computing DE for real data
91
+ INFO:cell_eval._evaluator:Using the following pdex kwargs: {'reference': 'control', 'groupby_key': 'perturbation', 'num_workers': 32, 'batch_size': 100, 'metric': 'wilcoxon', 'is_log1p': True, 'as_polars': True}
92
+ INFO:pdex._single_cell:Log1p status: True
93
+ INFO:pdex._single_cell:Precomputing masks for each target gene
94
+
95
+ INFO:pdex._single_cell:Precomputing variable indices for each feature
96
+
97
+ INFO:pdex._single_cell:Creating shared memory matrix for parallel computing
98
+ INFO:pdex._single_cell:Creating generator of all combinations: N=40000
99
+ INFO:pdex._single_cell:Creating generator of all batches: N=401
100
+ INFO:pdex._single_cell:Initializing parallel processing pool with 32 workers
101
+ INFO:pdex._single_cell:Processing batches
102
+
103
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
104
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
105
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
106
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
107
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
108
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
109
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
110
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
111
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
112
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
113
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
114
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
115
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
116
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
117
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
118
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
119
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
120
+
121
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
122
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
123
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
124
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
125
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
126
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
127
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
128
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
129
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
130
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
131
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
132
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
133
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
134
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
135
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
136
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
137
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
138
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
139
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
140
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
141
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
142
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
143
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
144
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
145
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
146
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
147
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
148
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
149
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
150
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
151
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
152
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
153
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
154
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
155
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
156
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
157
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
158
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
159
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
160
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
161
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
162
+
163
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
164
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
165
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
166
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
167
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
168
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
169
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
170
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
171
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
172
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
173
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
174
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
175
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
176
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
177
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
178
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
179
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
180
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
181
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
182
+
183
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
184
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
185
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
186
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
187
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
188
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
189
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
190
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
191
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
192
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
193
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
194
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
195
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
196
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
197
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
198
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
199
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
200
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
201
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
202
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
203
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
204
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
205
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
206
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
207
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
208
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
209
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
210
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
211
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
212
+
213
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
214
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
215
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
216
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
217
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
218
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
219
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
220
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
221
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
222
+
223
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
224
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
225
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
226
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
227
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
228
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
229
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
230
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
231
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
232
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
233
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
234
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
235
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
236
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
237
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
238
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
239
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
240
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
241
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
242
+
243
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
244
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
245
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
246
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
247
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
248
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
249
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
250
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
251
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
252
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
253
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
254
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
255
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
256
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
257
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
258
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
259
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
260
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
261
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
262
+
263
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
264
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
265
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
266
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
267
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
268
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
269
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
270
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
271
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
272
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
273
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
274
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
275
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
276
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
277
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
278
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
279
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
280
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
281
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
282
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
283
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
284
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
285
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
286
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
287
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
288
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
289
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
290
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
291
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
292
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
293
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
294
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
295
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
296
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
297
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
298
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
299
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
300
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
301
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
302
+
303
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
304
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
305
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
306
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
307
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
308
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
309
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
310
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
311
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
312
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
313
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
314
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
315
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
316
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
317
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
318
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
319
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
320
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
321
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
322
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
323
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
324
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
325
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
326
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
327
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
328
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
329
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
330
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
331
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
332
+
333
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
334
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
335
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
336
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
337
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
338
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
339
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
340
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
341
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
342
+
343
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
344
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
345
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
346
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
347
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
348
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
349
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
350
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
351
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
352
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
353
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
354
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
355
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
356
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
357
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
358
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
359
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
360
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scipy/stats/_mannwhitneyu.py:152: RuntimeWarning: invalid value encountered in sqrt
361
+ s = xp.sqrt(n1*n2/12 * ((n + 1) - tie_term/(n*(n-1))))
362
+
363
+ INFO:pdex._single_cell:Flattening results
364
+ INFO:pdex._single_cell:Closing shared memory pool
365
+ INFO:cell_eval._evaluator:Writing real DE results to: real_de.csv
366
+ INFO:cell_eval._evaluator:Computing DE for pred data
367
+ INFO:cell_eval._evaluator:Using the following pdex kwargs: {'reference': 'control', 'groupby_key': 'perturbation', 'num_workers': 32, 'batch_size': 100, 'metric': 'wilcoxon', 'is_log1p': True, 'as_polars': True}
368
+ INFO:pdex._single_cell:Log1p status: True
369
+ INFO:pdex._single_cell:Precomputing masks for each target gene
370
+
371
+ INFO:pdex._single_cell:Precomputing variable indices for each feature
372
+
373
+ INFO:pdex._single_cell:Creating shared memory matrix for parallel computing
374
+ INFO:pdex._single_cell:Creating generator of all combinations: N=40000
375
+ INFO:pdex._single_cell:Creating generator of all batches: N=401
376
+ INFO:pdex._single_cell:Initializing parallel processing pool with 32 workers
377
+ INFO:pdex._single_cell:Processing batches
378
+
379
+ INFO:pdex._single_cell:Flattening results
380
+ INFO:pdex._single_cell:Closing shared memory pool
381
+ INFO:cell_eval._evaluator:Writing pred DE results to: pred_de.csv
382
+ INFO:cell_eval._types._de:Checking DE data integrity... (real)
383
+ INFO:cell_eval._types._de:DE data integrity check complete. (real)
384
+ INFO:cell_eval._types._de:Checking DE data integrity... (pred)
385
+ INFO:cell_eval._types._de:DE data integrity check complete. (pred)
386
+ INFO:cell_eval._pipeline._runner:Computing metric 'overlap_at_N'
387
+ INFO:cell_eval._pipeline._runner:Computing metric 'overlap_at_50'
388
+ INFO:cell_eval._pipeline._runner:Computing metric 'overlap_at_100'
389
+ INFO:cell_eval._pipeline._runner:Computing metric 'overlap_at_200'
390
+ INFO:cell_eval._pipeline._runner:Computing metric 'overlap_at_500'
391
+ INFO:cell_eval._pipeline._runner:Computing metric 'precision_at_N'
392
+ INFO:cell_eval._pipeline._runner:Computing metric 'precision_at_50'
393
+ INFO:cell_eval._pipeline._runner:Computing metric 'precision_at_100'
394
+ INFO:cell_eval._pipeline._runner:Computing metric 'precision_at_200'
395
+ INFO:cell_eval._pipeline._runner:Computing metric 'precision_at_500'
396
+ INFO:cell_eval._pipeline._runner:Computing metric 'de_spearman_sig'
397
+ INFO:cell_eval._pipeline._runner:Computing metric 'de_direction_match'
398
+ INFO:cell_eval._pipeline._runner:Computing metric 'de_spearman_lfc_sig'
399
+ INFO:cell_eval._pipeline._runner:Computing metric 'de_sig_genes_recall'
400
+ INFO:cell_eval._pipeline._runner:Computing metric 'de_nsig_counts'
401
+ INFO:cell_eval._pipeline._runner:Computing metric 'pr_auc'
402
+ INFO:cell_eval._pipeline._runner:Computing metric 'roc_auc'
403
+ INFO:cell_eval._pipeline._runner:Computing metric 'pearson_delta'
404
+ INFO:cell_eval._types._anndata:Building pseudobulk embeddings for real anndata on: .X
405
+ INFO:cell_eval._types._anndata:Building pseudobulk embeddings for predicted anndata on: .X
406
+
407
+ INFO:cell_eval._pipeline._runner:Computing metric 'mse'
408
+
409
+ INFO:cell_eval._pipeline._runner:Computing metric 'mae'
410
+
411
+ INFO:cell_eval._pipeline._runner:Computing metric 'mse_delta'
412
+
413
+ INFO:cell_eval._pipeline._runner:Computing metric 'mae_delta'
414
+
415
+ INFO:cell_eval._pipeline._runner:Computing metric 'discrimination_score_l1'
416
+
417
+
418
+ INFO:cell_eval._pipeline._runner:Computing metric 'discrimination_score_l2'
419
+
420
+
421
+ INFO:cell_eval._pipeline._runner:Computing metric 'discrimination_score_cosine'
422
+
423
+
424
+ INFO:cell_eval._pipeline._runner:Computing metric 'pearson_edistance'
425
+ INFO:cell_eval.metrics._anndata:Precomputing sigma for control data (real)
426
+ INFO:cell_eval.metrics._anndata:Precomputing sigma for control data (pred)
427
+
428
+ INFO:cell_eval._pipeline._runner:Computing metric 'clustering_agreement'
429
+ INFO:cell_eval._evaluator:Writing perturbation level metrics to ./cell-eval-outdir/results.csv
430
+ INFO:cell_eval._evaluator:Writing aggregate metrics to ./cell-eval-outdir/agg_results.csv
431
+ ... storing 'perturbation' as categorical
432
+ ... storing 'perturbation' as categorical
433
+ Current evaluation score: 39.0000
434
+ Final evaluation score: 39.0000
435
+ ==========================================
436
+ Finished: Mon Mar 16 01:24:00 JST 2026
437
+ ==========================================
transfer/code/CCFM/logs/preextract_5402629.out ADDED
@@ -0,0 +1 @@
 
 
1
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/.d0005402629: line 13: /home/pj24002027/ku50002536/Takoai/lfj/stack_env/bin/activate: Permission denied
transfer/code/CCFM/logs/preextract_5402631.out ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==========================================
2
+ Job ID: 5402631
3
+ Job Name: preextract_scgpt
4
+ Start: Fri Mar 13 23:46:15 JST 2026
5
+ Node: b0034
6
+ GPU: NVIDIA H100, 95830 MiB
7
+ ==========================================
8
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scGPT/scgpt/model/model.py:21: UserWarning: flash_attn is not installed
9
+ warnings.warn("flash_attn is not installed")
10
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/scDFM/src/data_process/data.py:185: ImplicitModificationWarning: Trying to modify attribute `.obs` of view, initializing view as actual.
11
+ self.adata.obs['condition'] = self.adata.obs['condition'].str.replace('ctrl', 'control')
12
+ /home/hp250092/ku50001222/qian/aivc/lfj/stack_env/lib/python3.13/site-packages/scanpy/preprocessing/_highly_variable_genes.py:806: ImplicitModificationWarning: Trying to modify attribute `._uns` of view, initializing view as actual.
13
+ adata.uns["hvg"] = {"flavor": flavor}
14
+ /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/src/data/scgpt_extractor.py:109: UserWarning: FrozenScGPTExtractor: 498/5035 HVG genes not found in scGPT vocab, will use zero vectors.
15
+ warnings.warn(
16
+ Device: cuda
17
+ Converted var_names to gene symbols, sample: ['RP11-34P13.8', 'RP11-54O7.3', 'SAMD11', 'PERM1', 'HES4']
18
+ Warning: ctrl is not in the gene names
19
+ Cells: 91205, Genes: 5035
20
+ HVG gene names sample: ['RP11-34P13.8', 'RP11-54O7.3', 'SAMD11', 'PERM1', 'HES4']
21
+ Valid genes in scGPT vocab: 4537/5035, max_seq_len=4539
22
+ Loaded 159/163 pretrained parameters
23
+ scGPT d_model: 512
24
+ Output: /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scgpt_cache_norman.h5
25
+ Features shape: (91205, 5035, 512) float16
26
+
27
+ output = torch._nested_tensor_from_mask(
28
+
29
+ Global mean range: [-12.3974, 11.9278]
30
+ Global var range: [0.0192, 1.8779]
31
+ Done! Saved to /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/scgpt_cache_norman.h5
32
+ Features: (91205, 5035, 512) float16
33
+ Valid features counted: 413797085
34
+ ==========================================
35
+ Finished: Sat Mar 14 00:45:12 JST 2026
36
+ ==========================================
transfer/code/CCFM/plan.md ADDED
@@ -0,0 +1,490 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Cascaded Conditioned Flow Matching (CCFM) — 实现计划
2
+
3
+ ## Context
4
+
5
+ **目标**: 构建一个级联条件流匹配模型,用于单细胞扰动预测。以 scDFM 为底盘,引入 LatentForcing 的级联引导思想,使用 scGPT 提取的 **per-cell-per-gene contextualized features** 作为 latent 引导信号。
6
+
7
+ **核心概念映射**:
8
+ | LatentForcing (图像) | CCFM (单细胞) |
9
+ |---|---|
10
+ | Pixel patches → tokens | Gene expression → tokens |
11
+ | DINO-v2 feature patches → tokens | scGPT per-gene features → tokens |
12
+ | `pixel_embed + dino_embed` (element-wise add) | `expr_embed + scgpt_embed` (element-wise add) |
13
+ | `c = t_pixel + t_dino + y_emb` (AdaLN) | `c = t_expr + t_latent + pert_emb` (AdaLN) |
14
+ | Frozen DINO-v2 on-the-fly | Frozen scGPT on-the-fly |
15
+ | Separate pixel/dino decoder heads | Separate expr/latent decoder heads |
16
+
17
+ **数据集**: Norman (CRISPR) + ComboSciPlex (drug combination)
18
+
19
+ ---
20
+
21
+ ## 0. 项目结构
22
+
23
+ **工作目录**: `transfer/code/CCFM/`
24
+
25
+ ```
26
+ transfer/code/CCFM/
27
+ ├── plan.md # 本计划文件的副本
28
+ ├── config/
29
+ │ └── config_cascaded.py # CascadedFlowConfig 配置
30
+ ├── src/
31
+ │ ├── model/
32
+ │ │ ├── __init__.py
33
+ │ │ ├── model.py # CascadedFlowModel 主干网络
34
+ │ │ └── layers.py # 新增层: LatentEmbedder, LatentDecoder
35
+ │ ├── data/
36
+ │ │ ├── __init__.py
37
+ │ │ ├── data.py # 数据加载 (基于 scDFM, 新增 latent 支持)
38
+ │ │ └── scgpt_extractor.py # 冻结 scGPT 特征提取器封装
39
+ │ ├── denoiser.py # CascadedDenoiser (时间采样、损失、生成)
40
+ │ └── utils.py # 工具函数
41
+ ├── scripts/
42
+ │ ├── run_cascaded.py # 训练/推理入口
43
+ │ └── download_scgpt.py # 下载 scGPT 预训练模型
44
+ └── run.sh # pjsub 提交脚本模板
45
+ ```
46
+
47
+ 所有新代码写在 `transfer/code/CCFM/` 下,通过 import 引用 scDFM 和 scGPT 的现有模块。
48
+
49
+ ---
50
+
51
+ ## 1. 数据准备
52
+
53
+ ### 1.1 下载 scGPT 预训练模型
54
+
55
+ - 下载 `whole-human` 模型到 `transfer/data/scGPT_pretrained/`
56
+ - 包含: `best_model.pt`, `vocab.json`, `args.json`
57
+ - 脚本: `scripts/download_scgpt.py`
58
+
59
+ ### 1.2 数据加载 (`src/data/data.py`)
60
+
61
+ 复用 scDFM 的 `Data` 类进行数据加载和 train/test 划分。修改 `PerturbationDataset` 以返回 **raw expression values** (不仅是 gene indices),供 scGPT on-the-fly 提取使用。
62
+
63
+ 关键改动:
64
+ - 在 `__init__` 中预计算 scDFM gene names → scGPT vocab ID 的映射表
65
+ - `__getitem__` 返回: `src_cell_data`, `tgt_cell_data`, `condition_id`, `src_cell_raw` (供 scGPT), `tgt_cell_raw` (供 scGPT)
66
+
67
+ ### 1.3 scGPT 特征提取器 (`src/data/scgpt_extractor.py`)
68
+
69
+ **冻结 scGPT 作为 on-the-fly 特征提取器** (类似 LatentForcing 中冻结 DINO-v2):
70
+
71
+ ```python
72
+ class FrozenScGPTExtractor(nn.Module):
73
+ """
74
+ 封装冻结的 scGPT 模型,用于在训练时 on-the-fly 提取 per-gene features。
75
+ 类似 LatentForcing 的 dinov2_hf.py 中的 RAE 类。
76
+ """
77
+ def __init__(self, model_dir, n_hvg_genes, hvg_gene_names, device):
78
+ # 1. 加载预训练 scGPT (TransformerModel)
79
+ # 2. 冻结所有参数 (requires_grad=False)
80
+ # 3. 建立 hvg_gene_names → scGPT vocab ID 映射
81
+ # 4. 预计算全局归一化统计量 (running mean/var)
82
+
83
+ @torch.no_grad()
84
+ def extract(self, expression_values, gene_indices=None):
85
+ """
86
+ 输入: expression_values (B, G) — G=infer_top_gene 的表达值
87
+ 输出: per_gene_features (B, G, scgpt_d_model) — contextualized 特征
88
+
89
+ 流程:
90
+ 1. 筛选非零基因,构建 scGPT 输入 (token IDs + values)
91
+ 2. 运行冻结 scGPT._encode() → (B, seq_len, d_model)
92
+ 3. 将输出 scatter 回固定的 G 个基因位置
93
+ 4. 未覆盖的位置用零向量填充
94
+ 5. 全局归一化 + 方差匹配
95
+ """
96
+ ```
97
+
98
+ **关键技术细节**:
99
+ - scGPT 的 `_encode()` 返回 `(batch, seq_len, d_model=512)`,其中 seq_len 是变长的(仅包含非零表达的基因 + `<cls>` token)
100
+ - 需要用 gene ID 映射将变长输出 **scatter** 回固定的 G 个基因位置
101
+ - 基因对齐: scDFM 使用 gene symbol (如 "MAPK1"),scGPT 使用自己的 vocab。在 `__init__` 中建立 `hvg_name → scGPT_vocab_id` 的映射
102
+ - 不在 scGPT vocab 中的基因: 用零向量替代
103
+
104
+ ### 1.4 全局归一化与方差匹配 (CRITICAL)
105
+
106
+ 在 `FrozenScGPTExtractor` 内部实现:
107
+
108
+ ```python
109
+ # 方案: 使用 running statistics (类似 BatchNorm 的 running_mean/var)
110
+ # 在前 N 个 batch 中收集统计量,之后固定
111
+
112
+ # Step 1: 标准化到零均值单位方差
113
+ z_normalized = (z_raw - running_mean) / (running_std + 1e-6)
114
+
115
+ # Step 2: 方差匹配 — 缩放到与表达嵌入相近的量级
116
+ z_matched = z_normalized * target_std # target_std 默认 1.0
117
+ ```
118
+
119
+ 这对应 LatentForcing 中 `dinov2_hf.py` 的归一化:
120
+ ```python
121
+ z = (z - latent_mean) / sqrt(latent_var + eps) * match_pixel_norm
122
+ ```
123
+
124
+ ---
125
+
126
+ ## 2. 网络架构
127
+
128
+ ### 2.1 核心设计 — 严格参照 LatentForcing CoT (`model_cot.py`)
129
+
130
+ **LatentForcing 的做法 (model_cot.py:398-446)**:
131
+ 1. **双流嵌入 element-wise 相加** (line 414):
132
+ ```python
133
+ x = pixel_embedder(x_pixel) + dino_embedder(x_dino)
134
+ ```
135
+ 2. **条件向量求和通过 AdaLN** (line 409):
136
+ ```python
137
+ c = t_emb_pixel + t_emb_dino + y_emb
138
+ ```
139
+ 3. **共享 Transformer backbone**: 所有 block 处理融合后的 tokens
140
+ 4. **分离解码头** (line 429-441): `dh_blocks_pixel` + `dh_blocks_dino` 分别解码
141
+
142
+ **CCFM 的对应实现**:
143
+ 1. `x = expr_embedder(x_t_expr, gene_emb) + latent_embedder(z_t_latent)`
144
+ 2. `c = t_expr_emb + t_latent_emb + pert_emb`
145
+ 3. 复用 scDFM 的 DiffPerceiverBlock / PerceiverBlock
146
+ 4. ExprDecoder (复用) + LatentDecoder (新增)
147
+
148
+ ### 2.2 CascadedFlowModel 架构 (`src/model/model.py`)
149
+
150
+ ```
151
+ 输入:
152
+ - gene_id: (B, G) 基因 token IDs
153
+ - cell_1 (source): (B, G) control 细胞表达值
154
+ - x_t: (B, G) 加噪后的扰动细胞表达值 (expression flow)
155
+ - z_t: (B, G, 512) 加噪后的 scGPT per-gene 特征 (latent flow)
156
+ - t_expr: (B,) 表达流的时间步
157
+ - t_latent: (B,) 潜变量流的时间步
158
+ - perturbation_id: (B, 2) 扰动 token IDs
159
+
160
+ 数据流 (参照 model_cot.py 的 forward):
161
+ ┌──────────────────────────────────────────────────────┐
162
+ │ 1. 表达流嵌入 (复用 scDFM 的 origin 层) │
163
+ │ gene_emb = GeneEncoder(gene_id) (B,G,d) │
164
+ │ val_emb_1 = ValueEncoder_1(cell_1) + gene_emb │
165
+ │ val_emb_2 = ValueEncoder_2(x_t) + gene_emb │
166
+ │ expr_tokens = fusion_layer(cat(val_1, val_2)) │
167
+ │ → (B, G, d_model) │
168
+ ├──────────────────────────────────────────────────────┤
169
+ │ 2. Latent 流嵌入 (新增, 对应 dino_embedder) │
170
+ │ latent_tokens = LatentEmbedder(z_t) │
171
+ │ → (B, G, d_model) [bottleneck: 512 → d_model] │
172
+ ├──────────────────────────────────────────────────────┤
173
+ │ 3. Element-wise 相加 (对应 model_cot.py line 414) │
174
+ │ x = expr_tokens + latent_tokens │
175
+ │ → (B, G, d_model) │
176
+ ├──────────────────────────────────────────────────────┤
177
+ │ 4. 条件向量 (对应 model_cot.py line 409) │
178
+ │ c = t_expr_emb + t_latent_emb + pert_emb │
179
+ │ → (B, d_model) │
180
+ ├──────────────────────────────────────────────────────┤
181
+ │ 5. 共享 Backbone (复用 scDFM blocks) │
182
+ │ for i, block in enumerate(blocks): │
183
+ │ x = gene_adaLN[i](gene_emb, x) │
184
+ │ x = cat([x, pert_emb.expand], dim=-1) │
185
+ │ x = adapter_layer[i](x) │
186
+ │ x = block(x, val_emb_2, c) │
187
+ │ → (B, G, d_model) │
188
+ ├──────────────────────────────────────────────────────┤
189
+ │ 6a. 表达解码头 (复用 ExprDecoder) │
190
+ │ → pred_v_expr: (B, G) │
191
+ ├──────────────────────────────────────────────────────┤
192
+ │ 6b. Latent 解码头 (新增, 对应 dh_blocks_dino) │
193
+ │ latent_feats = dh_latent_proj(x) │
194
+ │ for block in dh_blocks_latent: │
195
+ │ latent_feats = block(latent_feats, c) │
196
+ │ pred_v_latent = latent_final_layer(latent_feats) │
197
+ │ → (B, G, 512) │
198
+ └──────────────────────────────────────────────────────┘
199
+
200
+ 输出: (pred_v_expr, pred_v_latent)
201
+ ```
202
+
203
+ ### 2.3 新增模块 (`src/model/layers.py`)
204
+
205
+ ```python
206
+ class LatentEmbedder(nn.Module):
207
+ """
208
+ 对应 LatentForcing 的 dino_embedder (BottleneckPatchEmbed)。
209
+ 将 per-gene scGPT 特征 (B, G, 512) 投影到 (B, G, d_model)。
210
+ """
211
+ def __init__(self, scgpt_dim=512, bottleneck_dim=128, d_model=512):
212
+ self.proj = nn.Sequential(
213
+ nn.Linear(scgpt_dim, bottleneck_dim),
214
+ nn.GELU(),
215
+ nn.Linear(bottleneck_dim, d_model),
216
+ )
217
+
218
+ def forward(self, z): # z: (B, G, scgpt_dim)
219
+ return self.proj(z) # (B, G, d_model)
220
+
221
+ class LatentDecoder(nn.Module):
222
+ """
223
+ 对应 LatentForcing 的 final_layer_dino + dh_blocks_dino。
224
+ 从 backbone 输出 (B, G, d_model) 解码回 (B, G, scgpt_dim)。
225
+ """
226
+ def __init__(self, d_model=512, scgpt_dim=512, dh_depth=2):
227
+ # 可选: 额外的 decoder head blocks
228
+ if dh_depth > 0:
229
+ self.dh_blocks = nn.ModuleList([...])
230
+ self.final = nn.Sequential(
231
+ nn.LayerNorm(d_model),
232
+ nn.Linear(d_model, d_model),
233
+ nn.GELU(),
234
+ nn.Linear(d_model, scgpt_dim),
235
+ )
236
+ ```
237
+
238
+ ### 2.4 复用的现有模块
239
+
240
+ | 模块 | 来源 | 文件路径 |
241
+ |------|------|---------|
242
+ | `TimestepEmbedder` | scDFM | `scDFM/src/models/origin/layers.py:157` |
243
+ | `ContinuousValueEncoder` | scDFM | `scDFM/src/models/origin/layers.py:26` |
244
+ | `GeneEncoder` | scDFM | `scDFM/src/models/origin/layers.py:55` |
245
+ | `BatchLabelEncoder` | scDFM | `scDFM/src/models/origin/layers.py:98` |
246
+ | `ExprDecoder` | scDFM | `scDFM/src/models/origin/layers.py:116` |
247
+ | `GeneadaLN` | scDFM | `scDFM/src/models/origin/layers.py:10` |
248
+ | `DiffPerceiverBlock` / `PerceiverBlock` | scDFM | `scDFM/src/models/origin/model.py:55,92` |
249
+ | `modulate` | scDFM | `scDFM/src/models/origin/blocks.py` |
250
+ | `AffineProbPath` + `CondOTScheduler` | scDFM | `scDFM/src/flow_matching/path/` |
251
+ | `make_lognorm_poisson_noise` | scDFM | `scDFM/src/utils/utils.py` |
252
+ | `TransformerModel` | scGPT | `scGPT/scgpt/model/model.py` |
253
+ | `GeneVocab` | scGPT | `scGPT/scgpt/tokenizer/gene_tokenizer.py` |
254
+ | `load_pretrained` | scGPT | `scGPT/scgpt/utils/util.py` |
255
+ | `DataCollator` | scGPT | `scGPT/scgpt/data_collator.py` |
256
+
257
+ ---
258
+
259
+ ## 3. 训练逻辑
260
+
261
+ ### 3.1 CascadedDenoiser (`src/denoiser.py`)
262
+
263
+ 封装 `CascadedFlowModel` + `FrozenScGPTExtractor`。
264
+
265
+ ### 3.2 级联时间步采样 — 参照 `denoiser_cot.py:121-132`
266
+
267
+ ```python
268
+ def sample_t(self, n, device):
269
+ """
270
+ dino_first_cascaded 模式 (denoiser_cot.py line 121-132):
271
+ - choose_latent_mask=True → 训练 latent 流
272
+ t_latent 随机, t_expr=0, loss_weight_expr=0
273
+ - choose_latent_mask=False → 训练 expression 流
274
+ t_expr 随机, t_latent=1 (clean latent), loss_weight_latent=0
275
+ """
276
+ t_latent = torch.rand(n, device=device)
277
+ t_expr = torch.rand(n, device=device)
278
+ choose_latent_mask = torch.rand(n, device=device) < self.choose_latent_p
279
+
280
+ t_latent = torch.where(choose_latent_mask, t_latent, torch.ones_like(t_latent))
281
+ t_expr = torch.where(choose_latent_mask, torch.zeros_like(t_expr), t_expr)
282
+
283
+ w_expr = (~choose_latent_mask).float()
284
+ w_latent = choose_latent_mask.float()
285
+ return t_expr, t_latent, w_expr, w_latent
286
+ ```
287
+
288
+ ### 3.3 训练步骤
289
+
290
+ ```python
291
+ def train_step(self, source, target, perturbation_id, gene_ids):
292
+ B = source.shape[0]
293
+
294
+ # 1. On-the-fly 提取 scGPT per-gene 特征 (冻结)
295
+ z_target = self.scgpt_extractor.extract(target) # (B, G, 512)
296
+
297
+ # 2. 时间采样
298
+ t_expr, t_latent, w_expr, w_latent = self.sample_t(B, device)
299
+
300
+ # 3. Expression path (同 scDFM)
301
+ noise_expr = make_noise(source, noise_type)
302
+ path_expr = affine_path.sample(t=t_expr, x_0=noise_expr, x_1=target)
303
+
304
+ # 4. Latent path
305
+ noise_latent = torch.randn_like(z_target)
306
+ path_latent = affine_path.sample(
307
+ t=t_latent.unsqueeze(-1).unsqueeze(-1), # broadcast to (B,1,1)
308
+ x_0=noise_latent, x_1=z_target
309
+ )
310
+
311
+ # 5. Model forward
312
+ pred_v_expr, pred_v_latent = model(
313
+ gene_ids, source, path_expr.x_t, path_latent.x_t,
314
+ t_expr, t_latent, perturbation_id
315
+ )
316
+
317
+ # 6. Losses
318
+ loss_expr = ((pred_v_expr - path_expr.dx_t)**2
319
+ * w_expr[:, None]).mean()
320
+ loss_latent = ((pred_v_latent - path_latent.dx_t)**2
321
+ * w_latent[:, None, None]).mean()
322
+ loss = loss_expr + latent_weight * loss_latent
323
+
324
+ # Optional MMD (同 scDFM)
325
+ if use_mmd_loss:
326
+ x1_hat = path_expr.x_t + pred_v_expr * (1 - t_expr).unsqueeze(-1)
327
+ loss += gamma * mmd_loss(x1_hat, target)
328
+
329
+ return loss
330
+ ```
331
+
332
+ ### 3.4 训练入口 (`scripts/run_cascaded.py`)
333
+
334
+ 基于 scDFM 的 `run.py` 改造:
335
+ - 初始化 `FrozenScGPTExtractor` + `CascadedFlowModel`
336
+ - 封装进 `CascadedDenoiser`
337
+ - 训练循环: 使用 `denoiser.train_step()` 替代 `train_step()`
338
+ - 评估: 使用 `denoiser.generate()` + `MetricsEvaluator`
339
+ - 其余(optimizer, scheduler, checkpoint)保持不变
340
+
341
+ ---
342
+
343
+ ## 4. 推理逻辑 — 两阶段级联生成
344
+
345
+ 参照 LatentForcing `denoiser_cot.py:224-247`:
346
+
347
+ ```python
348
+ @torch.no_grad()
349
+ def generate(self, source, perturbation_id, gene_ids,
350
+ latent_steps=20, expr_steps=20):
351
+ B = source.shape[0]
352
+ G = source.shape[1]
353
+
354
+ # ═══ Stage 1: 生成 Latent (per-gene scGPT features) ═══
355
+ # t_latent: 0→1, t_expr=0 (表达流不参与)
356
+ z_noise = torch.randn(B, G, scgpt_dim, device=device)
357
+
358
+ def latent_vf(t, z_t):
359
+ t_latent = t.expand(B)
360
+ t_expr = torch.zeros(B, device=device)
361
+ _, v_latent = model(gene_ids, source, source, z_t,
362
+ t_expr, t_latent, perturbation_id)
363
+ return v_latent
364
+
365
+ z_generated = torchdiffeq.odeint(
366
+ latent_vf, z_noise,
367
+ torch.linspace(0, 1, latent_steps, device=device),
368
+ method='rk4'
369
+ )[-1] # (B, G, 512)
370
+
371
+ # ═══ Stage 2: 用生成的 Latent 引导生成表达谱 ═══
372
+ # t_expr: 0→1, t_latent=1 (latent "clean")
373
+ expr_noise = make_noise(source, noise_type)
374
+
375
+ def expr_vf(t, x_t):
376
+ t_expr = t.expand(B)
377
+ t_latent = torch.ones(B, device=device)
378
+ v_expr, _ = model(gene_ids, source, x_t, z_generated,
379
+ t_expr, t_latent, perturbation_id)
380
+ return v_expr
381
+
382
+ x_generated = torchdiffeq.odeint(
383
+ expr_vf, expr_noise,
384
+ torch.linspace(0, 1, expr_steps, device=device),
385
+ method='rk4'
386
+ )[-1] # (B, G)
387
+
388
+ return torch.clamp(x_generated, min=0)
389
+ ```
390
+
391
+ **关键设计** (与 LatentForcing 的 `dino_first_cascaded` 推理完全对应):
392
+ - Stage 1: `t_expr=0` → 表达流为纯噪声,模型聚焦 latent 预测
393
+ - Stage 2: `t_latent=1` → latent 已解析,`z_generated` 通过 `LatentEmbedder` 投影后与表达 tokens 相加,引导表达生成
394
+ - 评估: 复用 scDFM 的 `test()` 函数结构 + `cell-eval MetricsEvaluator`
395
+
396
+ ---
397
+
398
+ ## 5. 配置 (`config/config_cascaded.py`)
399
+
400
+ ```python
401
+ @dataclass
402
+ class CascadedFlowConfig:
403
+ # === 基础 (同 scDFM FlowConfig) ===
404
+ model_type: str = 'cascaded'
405
+ batch_size: int = 48
406
+ d_model: int = 128
407
+ nhead: int = 8
408
+ nlayers: int = 4
409
+ lr: float = 5e-5
410
+ steps: int = 200000
411
+ eta_min: float = 1e-6
412
+ data_name: str = 'norman'
413
+ fusion_method: str = 'differential_perceiver'
414
+ perturbation_function: str = 'crisper'
415
+ noise_type: str = 'Gaussian'
416
+ infer_top_gene: int = 1000
417
+ n_top_genes: int = 5000
418
+ print_every: int = 5000
419
+ gamma: float = 0.5
420
+ use_mmd_loss: bool = True
421
+ split_method: str = 'additive'
422
+ fold: int = 1
423
+
424
+ # === 级联/Latent 特有 ===
425
+ scgpt_dim: int = 512 # scGPT embedding 维度
426
+ bottleneck_dim: int = 128 # LatentEmbedder 瓶颈维度
427
+ latent_weight: float = 1.0 # latent 流损失权重
428
+ choose_latent_p: float = 0.4 # 训练 latent 流的概率
429
+ target_std: float = 1.0 # 方差匹配目标标准差
430
+ dh_depth: int = 2 # latent decoder head 层数
431
+
432
+ # === scGPT 路径 ===
433
+ scgpt_model_dir: str = 'transfer/data/scGPT_pretrained'
434
+
435
+ # === 推理 ===
436
+ latent_steps: int = 20
437
+ expr_steps: int = 20
438
+ ```
439
+
440
+ ---
441
+
442
+ ## 6. 实现顺序
443
+
444
+ ### Step 1: 项目初始化 + 下载 scGPT
445
+ - 创建 `transfer/code/CCFM/` 目录结构
446
+ - 下载 scGPT `whole-human` 预训练模型
447
+ - 验证: 加载 scGPT 模型成功
448
+
449
+ ### Step 2: FrozenScGPTExtractor
450
+ - 实现 `src/data/scgpt_extractor.py`
451
+ - 处理 scDFM ↔ scGPT 的基因名对齐
452
+ - 实现 per-gene feature scatter 和归一化
453
+ - 验证: 输入 (48, 1000) expression → 输出 (48, 1000, 512) features
454
+
455
+ ### Step 3: 数据加载
456
+ - 实现 `src/data/data.py`,复用 scDFM 的 Data 类
457
+ - 验证: DataLoader 正确返回 expression + perturbation_id
458
+
459
+ ### Step 4: CascadedFlowModel
460
+ - 实现 `src/model/model.py` 和 `src/model/layers.py`
461
+ - 验证: dummy forward pass → (pred_v_expr, pred_v_latent) 形状正确
462
+
463
+ ### Step 5: CascadedDenoiser
464
+ - 实现 `src/denoiser.py`
465
+ - 验证: `train_step()` 返回 scalar loss, `generate()` 返回正确形状
466
+
467
+ ### Step 6: 训练入口 + 配置
468
+ - 实现 `scripts/run_cascaded.py` + `config/config_cascaded.py`
469
+ - 编写 `run.sh` (pjsub 脚本)
470
+ - 验证: 完整训练循环运行,loss 下降
471
+
472
+ ### Step 7: 端到端验证
473
+ - Norman 数据集上训练(先小 step 数)
474
+ - 检查 loss_expr 和 loss_latent 分别下降
475
+ - 推理并通过 cell-eval 评估
476
+ - 与 baseline scDFM 对比
477
+
478
+ ---
479
+
480
+ ## 7. 关键技术决策
481
+
482
+ | 决策点 | 选择 | 理由 |
483
+ |--------|------|------|
484
+ | Latent 注入方式 | Element-wise add (token-level) | 严格对标 LatentForcing `model_cot.py:414` |
485
+ | 条件注入 | AdaLN (c = sum of all embeddings) | 对标 `model_cot.py:409` |
486
+ | scGPT 特征提取 | On-the-fly (冻结) | 存储不可行 (~1TB); 对标 LatentForcing 冻结 DINO-v2 |
487
+ | 时间步采样 | dino_first_cascaded | 对标 `denoiser_cot.py:121-132` |
488
+ | 方差匹配 | 全局 running statistics + scaling | 对标 `dinov2_hf.py` 的归一化 |
489
+ | Backbone | 复用 scDFM 的 DiffPerceiver/Perceiver | 保持底盘一致 |
490
+ | 解码头 | 分离: ExprDecoder + LatentDecoder | 对标 `model_cot.py:429-441` 的分离头 |
transfer/code/CCFM/preextract_scgpt.5402631.stats ADDED
@@ -0,0 +1,467 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Job Statistical Information
2
+
3
+ JOB ID : 5402631
4
+ SUB JOB NUM : -
5
+ START BULKNO : -
6
+ END BULKNO : -
7
+ HOST NAME : genkai0002
8
+ JOB NAME : preextract_scgpt
9
+ JOB TYPE : BATCH
10
+ JOB MODEL : NM
11
+ USER : ku50001222
12
+ GROUP : hp250092
13
+ RESOURCE UNIT : rscunit_pg01 <DEFAULT>
14
+ RESOURCE GROUP : b-batch
15
+ APRIORITY : 127 <DEFAULT>
16
+ PRIORITY : 127 <DEFAULT>
17
+ SHELL : /bin/bash
18
+ MAIL SEND FLAG : ----
19
+ MAIL ADDRESS : ku50001222
20
+ STEP DEPENDENCY EXP :
21
+ FILE MASK : 0022
22
+ STANDARD OUT FILE : /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/logs/preextract_5402631.out
23
+ STANDARD ERR FILE : /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/logs/preextract_5402631.out
24
+ INFORMATION FILE : /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM/preextract_scgpt.5402631.stats
25
+ PJSUB DIRECTORY : /home/hp250092/ku50001222/qian/aivc/lfj/transfer/code/CCFM
26
+ FILE SYSTEM NAME :
27
+ APPLICATION NAME :
28
+ NUMA POLICY : 0
29
+ SPECIFIED JOB START DATE : -
30
+ AUTO RESTART : yes
31
+ MPI PROC (REQUIRE) : 1
32
+ MPI PROC (ALLOC) : 1
33
+ MPI STATIC PROC (REQUIRE) : 1
34
+ MPI STATIC PROC (ALLOC) : 1
35
+ PROC CPU LIMIT : unlimited <DEFAULT>
36
+ PROC COREFILE LIMIT : 0.0 MB (0) <DEFAULT>
37
+ PROC CREATE PROCESS LIMIT : 16384 <DEFAULT>
38
+ PROC DATA LIMIT : unlimited <DEFAULT>
39
+ PROC LOCKED MEMORY LIMIT : unlimited <DEFAULT>
40
+ PROC MESSAGE QUEUE LIMIT : unlimited <DEFAULT>
41
+ PROC OPEN FILES LIMIT : 1048576 <DEFAULT>
42
+ PROC PENDING SIGNAL LIMIT : unlimited <DEFAULT>
43
+ PROC PERMFILE LIMIT : unlimited <DEFAULT>
44
+ PROC STACK LIMIT : unlimited <DEFAULT>
45
+ PROC VIRTUAL MEMORY LIMIT : unlimited <DEFAULT>
46
+ COMMENT :
47
+ SUPPLEMENTARY INFORMATION :
48
+ RETRY NUM : 0
49
+ LAST STATE : RNO
50
+ STATE : EXT
51
+ BACKFILL : 1
52
+ ACCEPT DATE : 2026/03/13 23:46:01
53
+ QUEUED DATE : 2026/03/13 23:46:01
54
+ PRO START DATE : -
55
+ PRO END DATE : -
56
+ JOB START DATE : 2026/03/13 23:46:14
57
+ JOB END DATE : 2026/03/14 00:45:13
58
+ EPI START DATE : -
59
+ EPI END DATE : -
60
+ EXIT DATE : 2026/03/14 00:45:14
61
+ JOB DELETE DATE (REQUIRE) : -
62
+ JOB DELETE DATE : -
63
+ ALL PREC SUBJOB EXIT DATE : -
64
+ HOLD TIME : 0
65
+ LAST HOLD USER : -
66
+ HOLD NUM : 0
67
+ PRM DATE : 2026/03/14 00:45:13
68
+ AFFECTED NODE ID : -
69
+ EXIT CODE : 0
70
+ SIGNAL NO : -
71
+ PJM CODE : 0
72
+ REASON : -
73
+ PRO EXIT CODE : -
74
+ EPI EXIT CODE : -
75
+ ELAPSE TIME (LIMIT) : 03:00:00 (10800)
76
+ ELAPSE TIME (USE) : 00:58:59 (3539)
77
+ USER CPU TIME (USE) : 5779310 ms
78
+ SYSTEM CPU TIME (USE) : 3981010 ms
79
+ CPU TIME (TOTAL) : 9760320 ms
80
+ MEMORY SIZE (LIMIT) : -
81
+ MEMORY SIZE (ALLOC) : -
82
+ MAX MEMORY SIZE (USE) : -
83
+ VNODE MEMORY SIZE (LIMIT) : 232050.0 MiB (243322060800)
84
+ VNODE MEMORY SIZE (ALLOC) : 232050.0 MiB (243322060800)
85
+ VNODE MAX MEMORY SIZE (USE) : 232049.3 MiB (243321311232)
86
+ CPU NUM (REQUIRE) : 30
87
+ CPU NUM (ALLOC) : 30
88
+ CPU NUM (USE) : 30
89
+ TOFU USER COMM RECV BYTE : -
90
+ TOFU USER COMM SEND BYTE : -
91
+ TOFU SYSTEM COMM RECV BYTE : -
92
+ TOFU SYSTEM COMM SEND BYTE : -
93
+ FJ PROFILER : -
94
+ SECTOR CACHE : -
95
+ INTRA NODE BARRIER : -
96
+ AVG POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
97
+ MAX POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
98
+ MIN POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
99
+ ENERGY CONSUMPTION OF CORES/CMG (IDEAL) : -
100
+ AVG POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
101
+ MAX POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
102
+ MIN POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
103
+ ENERGY CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
104
+ AVG POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
105
+ MAX POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
106
+ MIN POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
107
+ ENERGY CONSUMPTION OF MEM/CMG (IDEAL) : -
108
+ AVG POWER CONSUMPTION OF TOFU (IDEAL) : -
109
+ MAX POWER CONSUMPTION OF TOFU (IDEAL) : -
110
+ MIN POWER CONSUMPTION OF TOFU (IDEAL) : -
111
+ ENERGY CONSUMPTION OF TOFU (IDEAL) : -
112
+ AVG POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
113
+ MAX POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
114
+ MIN POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
115
+ ENERGY CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
116
+ AVG POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
117
+ MAX POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
118
+ MIN POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
119
+ ENERGY CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
120
+ AVG POWER CONSUMPTION OF PCI-E (IDEAL) : -
121
+ MAX POWER CONSUMPTION OF PCI-E (IDEAL) : -
122
+ MIN POWER CONSUMPTION OF PCI-E (IDEAL) : -
123
+ ENERGY CONSUMPTION OF PCI-E (IDEAL) : -
124
+ AVG POWER CONSUMPTION OF NODE (IDEAL) : -
125
+ MAX POWER CONSUMPTION OF NODE (IDEAL) : -
126
+ MIN POWER CONSUMPTION OF NODE (IDEAL) : -
127
+ ENERGY CONSUMPTION OF NODE (IDEAL) : -
128
+ AVG POWER CONSUMPTION OF NODE (MEASURED) : -
129
+ MAX POWER CONSUMPTION OF NODE (MEASURED) : -
130
+ MIN POWER CONSUMPTION OF NODE (MEASURED) : -
131
+ ENERGY CONSUMPTION OF NODE (MEASURED) : -
132
+ UTILIZATION INFO OF POWER API : 0
133
+ AVG POWER CONSUMPTION OF CPU/PKG : -
134
+ MAX POWER CONSUMPTION OF CPU/PKG : -
135
+ MIN POWER CONSUMPTION OF CPU/PKG : -
136
+ ENERGY CONSUMPTION OF CPU/PKG : -
137
+ AVG POWER CONSUMPTION OF MEM/PKG : -
138
+ MAX POWER CONSUMPTION OF MEM/PKG : -
139
+ MIN POWER CONSUMPTION OF MEM/PKG : -
140
+ ENERGY CONSUMPTION OF MEM/PKG : -
141
+ AVG POWER CONSUMPTION OF PP0/PKG : -
142
+ MAX POWER CONSUMPTION OF PP0/PKG : -
143
+ MIN POWER CONSUMPTION OF PP0/PKG : -
144
+ ENERGY CONSUMPTION OF PP0/PKG : -
145
+ POWER CONSUMPTION STATE : 4
146
+ POWER CONSUMPTION MEASURE START DATE : -
147
+ POWER CONSUMPTION MEASURE END DATE : -
148
+ VNODE NUM (REQUIRE) : 1
149
+ VNODE NUM (ALLOC) : 1
150
+ VNODE NUM (USE) : 1
151
+ NODE NUM (REQUIRE) : -
152
+ NODE NUM (ALLOC) : -
153
+ NODE NUM (USE) : -
154
+ NODE NUM (UNUSED) : 0
155
+ NODE ID (USE) : 0x02FF0024
156
+ TOFU COORDINATE (USE) : -
157
+ NODE ID (RANK) : 0x02FF0024(0)
158
+ jobenv (REQUIRE) :
159
+ jobenv (ALLOC) :
160
+ jobenv (USE) :
161
+ gpu (REQUIRE) : 1
162
+ gpu (ALLOC) : 1
163
+ gpu (USE) : -
164
+ simplex (REQUIRE) :
165
+ simplex (ALLOC) :
166
+ simplex (USE) :
167
+ shared (REQUIRE) : true
168
+ shared (ALLOC) : true
169
+ shared (USE) :
170
+ short-job (REQUIRE) : false
171
+ short-job (ALLOC) : false
172
+ short-job (USE) :
173
+ rsc001 (REQUIRE) : 0 <DEFAULT>
174
+ rsc001 (ALLOC) : 0
175
+ rsc001 (USE) : -
176
+ rsc002 (REQUIRE) : 0 <DEFAULT>
177
+ rsc002 (ALLOC) : 0
178
+ rsc002 (USE) : -
179
+ rsc003 (REQUIRE) : 0 <DEFAULT>
180
+ rsc003 (ALLOC) : 0
181
+ rsc003 (USE) : -
182
+ rsc004 (REQUIRE) : 0 <DEFAULT>
183
+ rsc004 (ALLOC) : 0
184
+ rsc004 (USE) : -
185
+ rsc005 (REQUIRE) : 0 <DEFAULT>
186
+ rsc005 (ALLOC) : 0
187
+ rsc005 (USE) : -
188
+ rsc006 (REQUIRE) : 0 <DEFAULT>
189
+ rsc006 (ALLOC) : 0
190
+ rsc006 (USE) : -
191
+ rsc007 (REQUIRE) : 0 <DEFAULT>
192
+ rsc007 (ALLOC) : 0
193
+ rsc007 (USE) : -
194
+ rsc008 (REQUIRE) : 0 <DEFAULT>
195
+ rsc008 (ALLOC) : 0
196
+ rsc008 (USE) : -
197
+ rsc009 (REQUIRE) : 0 <DEFAULT>
198
+ rsc009 (ALLOC) : 0
199
+ rsc009 (USE) : -
200
+ rsc010 (REQUIRE) : 0 <DEFAULT>
201
+ rsc010 (ALLOC) : 0
202
+ rsc010 (USE) : -
203
+ rsc011 (REQUIRE) : 0 <DEFAULT>
204
+ rsc011 (ALLOC) : 0
205
+ rsc011 (USE) : -
206
+ rsc012 (REQUIRE) : 0 <DEFAULT>
207
+ rsc012 (ALLOC) : 0
208
+ rsc012 (USE) : -
209
+ rsc013 (REQUIRE) : 0 <DEFAULT>
210
+ rsc013 (ALLOC) : 0
211
+ rsc013 (USE) : -
212
+ rsc014 (REQUIRE) : 0 <DEFAULT>
213
+ rsc014 (ALLOC) : 0
214
+ rsc014 (USE) : -
215
+ rsc015 (REQUIRE) : 0 <DEFAULT>
216
+ rsc015 (ALLOC) : 0
217
+ rsc015 (USE) : -
218
+ rsc016 (REQUIRE) : 0 <DEFAULT>
219
+ rsc016 (ALLOC) : 0
220
+ rsc016 (USE) : -
221
+ rsc017 (REQUIRE) : 0 <DEFAULT>
222
+ rsc017 (ALLOC) : 0
223
+ rsc017 (USE) : -
224
+ rsc018 (REQUIRE) : 0 <DEFAULT>
225
+ rsc018 (ALLOC) : 0
226
+ rsc018 (USE) : -
227
+ rsc019 (REQUIRE) : 0 <DEFAULT>
228
+ rsc019 (ALLOC) : 0
229
+ rsc019 (USE) : -
230
+ rsc020 (REQUIRE) : 0 <DEFAULT>
231
+ rsc020 (ALLOC) : 0
232
+ rsc020 (USE) : -
233
+ rsc021 (REQUIRE) : 0 <DEFAULT>
234
+ rsc021 (ALLOC) : 0
235
+ rsc021 (USE) : -
236
+ rsc022 (REQUIRE) : 0 <DEFAULT>
237
+ rsc022 (ALLOC) : 0
238
+ rsc022 (USE) : -
239
+ rsc023 (REQUIRE) : 0 <DEFAULT>
240
+ rsc023 (ALLOC) : 0
241
+ rsc023 (USE) : -
242
+ rsc024 (REQUIRE) : 0 <DEFAULT>
243
+ rsc024 (ALLOC) : 0
244
+ rsc024 (USE) : -
245
+ rsc025 (REQUIRE) : 0 <DEFAULT>
246
+ rsc025 (ALLOC) : 0
247
+ rsc025 (USE) : -
248
+ rsc026 (REQUIRE) : 0 <DEFAULT>
249
+ rsc026 (ALLOC) : 0
250
+ rsc026 (USE) : -
251
+ rsc027 (REQUIRE) : 0 <DEFAULT>
252
+ rsc027 (ALLOC) : 0
253
+ rsc027 (USE) : -
254
+ rsc028 (REQUIRE) : 0 <DEFAULT>
255
+ rsc028 (ALLOC) : 0
256
+ rsc028 (USE) : -
257
+ rsc029 (REQUIRE) : 0 <DEFAULT>
258
+ rsc029 (ALLOC) : 0
259
+ rsc029 (USE) : -
260
+ rsc030 (REQUIRE) : 0 <DEFAULT>
261
+ rsc030 (ALLOC) : 0
262
+ rsc030 (USE) : -
263
+ rsc031 (REQUIRE) : 0 <DEFAULT>
264
+ rsc031 (ALLOC) : 0
265
+ rsc031 (USE) : -
266
+ rsc032 (REQUIRE) : 0 <DEFAULT>
267
+ rsc032 (ALLOC) : 0
268
+ rsc032 (USE) : -
269
+ rsc033 (REQUIRE) : 0 <DEFAULT>
270
+ rsc033 (ALLOC) : 0
271
+ rsc033 (USE) : -
272
+ rsc034 (REQUIRE) : 0 <DEFAULT>
273
+ rsc034 (ALLOC) : 0
274
+ rsc034 (USE) : -
275
+ rsc035 (REQUIRE) : 0 <DEFAULT>
276
+ rsc035 (ALLOC) : 0
277
+ rsc035 (USE) : -
278
+ rsc036 (REQUIRE) : 0 <DEFAULT>
279
+ rsc036 (ALLOC) : 0
280
+ rsc036 (USE) : -
281
+ rsc037 (REQUIRE) : 0 <DEFAULT>
282
+ rsc037 (ALLOC) : 0
283
+ rsc037 (USE) : -
284
+ rsc038 (REQUIRE) : 0 <DEFAULT>
285
+ rsc038 (ALLOC) : 0
286
+ rsc038 (USE) : -
287
+ rsc039 (REQUIRE) : 0 <DEFAULT>
288
+ rsc039 (ALLOC) : 0
289
+ rsc039 (USE) : -
290
+ rsc040 (REQUIRE) : 0 <DEFAULT>
291
+ rsc040 (ALLOC) : 0
292
+ rsc040 (USE) : -
293
+ rsc041 (REQUIRE) : 0 <DEFAULT>
294
+ rsc041 (ALLOC) : 0
295
+ rsc041 (USE) : -
296
+ rsc042 (REQUIRE) : 0 <DEFAULT>
297
+ rsc042 (ALLOC) : 0
298
+ rsc042 (USE) : -
299
+ rsc043 (REQUIRE) : 0 <DEFAULT>
300
+ rsc043 (ALLOC) : 0
301
+ rsc043 (USE) : -
302
+ rsc044 (REQUIRE) : 0 <DEFAULT>
303
+ rsc044 (ALLOC) : 0
304
+ rsc044 (USE) : -
305
+ rsc045 (REQUIRE) : 0 <DEFAULT>
306
+ rsc045 (ALLOC) : 0
307
+ rsc045 (USE) : -
308
+ rsc046 (REQUIRE) : 0 <DEFAULT>
309
+ rsc046 (ALLOC) : 0
310
+ rsc046 (USE) : -
311
+ rsc047 (REQUIRE) : 0 <DEFAULT>
312
+ rsc047 (ALLOC) : 0
313
+ rsc047 (USE) : -
314
+ rsc048 (REQUIRE) : 0 <DEFAULT>
315
+ rsc048 (ALLOC) : 0
316
+ rsc048 (USE) : -
317
+ rsc049 (REQUIRE) : 0 <DEFAULT>
318
+ rsc049 (ALLOC) : 0
319
+ rsc049 (USE) : -
320
+ rsc050 (REQUIRE) : 0 <DEFAULT>
321
+ rsc050 (ALLOC) : 0
322
+ rsc050 (USE) : -
323
+
324
+ --------------------------------------------------------------------------------
325
+ Node Statistical Information
326
+
327
+ VNODE ID : 0
328
+ NODE ID : 0x02FF0024
329
+ TOFU COORDINATE : -
330
+ NODE COORDINATE : -
331
+ RANK NO : 0
332
+ NODE AVAILABLE : available
333
+ ARCHI INFORMATION : PG
334
+ CPU BITMAP (ALLOC) : 0xfffffffc00000000000000000000000
335
+ CPU BITMAP (USE) : 0xfffffffc00000000000000000000000
336
+ MPI PROC (ALLOC) : 1
337
+ MPI STATIC PROC (ALLOC) : 1
338
+ USER CPU TIME (USE) : 5779310 ms
339
+ SYSTEM CPU TIME (USE) : 3981010 ms
340
+ CPU TIME (TOTAL) : 9760320 ms
341
+ MEMORY SIZE (ALLOC) : 232050.0 MiB (243322060800)
342
+ MAX MEMORY SIZE (USE) : 232049.3 MiB (243321311232)
343
+ CPU NUM (ALLOC) : 30
344
+ CPU NUM (USE) : 30
345
+ PERF COUNT 1 : -
346
+ PERF COUNT 2 : -
347
+ PERF COUNT 3 : -
348
+ PERF COUNT 4 : -
349
+ PERF COUNT 5 : -
350
+ PERF COUNT 6 : -
351
+ PERF COUNT 7 : -
352
+ PERF COUNT 8 : -
353
+ PERF COUNT 9 : -
354
+ TOFU USER COMM RECV BYTE : -
355
+ TOFU USER COMM SEND BYTE : -
356
+ TOFU SYSTEM COMM RECV BYTE : -
357
+ TOFU SYSTEM COMM SEND BYTE : -
358
+ FJ PROFILER : -
359
+ SECTOR CACHE : -
360
+ INTRA NODE BARRIER : -
361
+ AVG POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
362
+ MAX POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
363
+ MIN POWER CONSUMPTION OF CORES/CMG (IDEAL) : -
364
+ ENERGY CONSUMPTION OF CORES/CMG (IDEAL) : -
365
+ AVG POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
366
+ MAX POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
367
+ MIN POWER CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
368
+ ENERGY CONSUMPTION OF L2CACHE/CMG (IDEAL) : -
369
+ AVG POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
370
+ MAX POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
371
+ MIN POWER CONSUMPTION OF MEM/CMG (IDEAL) : -
372
+ ENERGY CONSUMPTION OF MEM/CMG (IDEAL) : -
373
+ AVG POWER CONSUMPTION OF TOFU (IDEAL) : -
374
+ MAX POWER CONSUMPTION OF TOFU (IDEAL) : -
375
+ MIN POWER CONSUMPTION OF TOFU (IDEAL) : -
376
+ ENERGY CONSUMPTION OF TOFU (IDEAL) : -
377
+ AVG POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
378
+ MAX POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
379
+ MIN POWER CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
380
+ ENERGY CONSUMPTION OF CPU PERIPHERALS (IDEAL) : -
381
+ AVG POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
382
+ MAX POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
383
+ MIN POWER CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
384
+ ENERGY CONSUMPTION OF OPTICAL MODULE (IDEAL) : -
385
+ AVG POWER CONSUMPTION OF PCI-E (IDEAL) : -
386
+ MAX POWER CONSUMPTION OF PCI-E (IDEAL) : -
387
+ MIN POWER CONSUMPTION OF PCI-E (IDEAL) : -
388
+ ENERGY CONSUMPTION OF PCI-E (IDEAL) : -
389
+ AVG POWER CONSUMPTION OF NODE (IDEAL) : -
390
+ MAX POWER CONSUMPTION OF NODE (IDEAL) : -
391
+ MIN POWER CONSUMPTION OF NODE (IDEAL) : -
392
+ ENERGY CONSUMPTION OF NODE (IDEAL) : -
393
+ AVG POWER CONSUMPTION OF NODE (MEASURED) : -
394
+ MAX POWER CONSUMPTION OF NODE (MEASURED) : -
395
+ MIN POWER CONSUMPTION OF NODE (MEASURED) : -
396
+ ENERGY CONSUMPTION OF NODE (MEASURED) : -
397
+ UTILIZATION INFO OF POWER API : 0
398
+ AVG POWER CONSUMPTION OF CPU/PKG : -
399
+ MAX POWER CONSUMPTION OF CPU/PKG : -
400
+ MIN POWER CONSUMPTION OF CPU/PKG : -
401
+ ENERGY CONSUMPTION OF CPU/PKG : -
402
+ AVG POWER CONSUMPTION OF MEM/PKG : -
403
+ MAX POWER CONSUMPTION OF MEM/PKG : -
404
+ MIN POWER CONSUMPTION OF MEM/PKG : -
405
+ ENERGY CONSUMPTION OF MEM/PKG : -
406
+ AVG POWER CONSUMPTION OF PP0/PKG : -
407
+ MAX POWER CONSUMPTION OF PP0/PKG : -
408
+ MIN POWER CONSUMPTION OF PP0/PKG : -
409
+ ENERGY CONSUMPTION OF PP0/PKG : -
410
+ POWER CONSUMPTION STATE : 4
411
+ POWER CONSUMPTION MEASURE START DATE : -
412
+ POWER CONSUMPTION MEASURE END DATE : -
413
+ gpu (REQUIRE) : 1
414
+ gpu (ALLOC) : 1
415
+ gpu (USE) : -
416
+ simplex (REQUIRE) :
417
+ simplex (ALLOC) :
418
+ simplex (USE) :
419
+ shared (REQUIRE) : true
420
+ shared (ALLOC) : true
421
+ shared (USE) :
422
+ short-job (REQUIRE) : false
423
+ short-job (ALLOC) : false
424
+ short-job (USE) :
425
+ rsc001 (REQUIRE) : 0 <DEFAULT>
426
+ rsc001 (ALLOC) : 0
427
+ rsc001 (USE) : -
428
+ rsc002 (REQUIRE) : 0 <DEFAULT>
429
+ rsc002 (ALLOC) : 0
430
+ rsc002 (USE) : -
431
+ rsc003 (REQUIRE) : 0 <DEFAULT>
432
+ rsc003 (ALLOC) : 0
433
+ rsc003 (USE) : -
434
+ rsc004 (REQUIRE) : 0 <DEFAULT>
435
+ rsc004 (ALLOC) : 0
436
+ rsc004 (USE) : -
437
+ rsc005 (REQUIRE) : 0 <DEFAULT>
438
+ rsc005 (ALLOC) : 0
439
+ rsc005 (USE) : -
440
+ rsc006 (REQUIRE) : 0 <DEFAULT>
441
+ rsc006 (ALLOC) : 0
442
+ rsc006 (USE) : -
443
+ rsc007 (REQUIRE) : 0 <DEFAULT>
444
+ rsc007 (ALLOC) : 0
445
+ rsc007 (USE) : -
446
+ rsc008 (REQUIRE) : 0 <DEFAULT>
447
+ rsc008 (ALLOC) : 0
448
+ rsc008 (USE) : -
449
+ rsc009 (REQUIRE) : 0 <DEFAULT>
450
+ rsc009 (ALLOC) : 0
451
+ rsc009 (USE) : -
452
+ rsc010 (REQUIRE) : 0 <DEFAULT>
453
+ rsc010 (ALLOC) : 0
454
+ rsc010 (USE) : -
455
+ rsc011 (REQUIRE) : 0 <DEFAULT>
456
+ rsc011 (ALLOC) : 0
457
+ rsc011 (USE) : -
458
+ rsc012 (REQUIRE) : 0 <DEFAULT>
459
+ rsc012 (ALLOC) : 0
460
+ rsc012 (USE) : -
461
+ rsc013 (REQUIRE) : 0 <DEFAULT>
462
+ rsc013 (ALLOC) : 0
463
+ rsc013 (USE) : -
464
+ rsc014 (REQUIRE) : 0 <DEFAULT>
465
+ rsc014 (ALLOC) : 0
466
+ rsc014 (USE) : -
467
+
transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/agg_results.csv ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ statistic,overlap_at_N,overlap_at_50,overlap_at_100,overlap_at_200,overlap_at_500,precision_at_N,precision_at_50,precision_at_100,precision_at_200,precision_at_500,de_spearman_sig,de_direction_match,de_spearman_lfc_sig,de_sig_genes_recall,de_nsig_counts_real,de_nsig_counts_pred,pr_auc,roc_auc,pearson_delta,mse,mae,mse_delta,mae_delta,discrimination_score_l1,discrimination_score_l2,discrimination_score_cosine,pearson_edistance,clustering_agreement
2
+ count,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0
3
+ null_count,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
4
+ mean,0.07343027035430702,0.06387820512820516,0.07076309811824325,0.07205733118513703,0.07343027035430702,0.13271404481929105,0.06410256410256414,0.07230769230769234,0.07051282051282053,0.08923076923076927,-0.0009229119228736443,0.7554765592002619,0.7265820647068527,0.9465994535977555,137.89743589743588,987.5128205128206,0.20047403116204476,0.5081252739494853,-0.005123983103770149,0.1430895901787359,0.16863651470832744,0.1430895901787359,0.1686365147083274,0.5259697567389875,0.5180802103879028,0.5174227481919791,0.011980186563896652,0.023888740241565534
5
+ std,0.060600501863245414,0.04629424881137447,0.051792809576235235,0.05641536129874057,0.060600501863245414,0.07196063490463939,0.046152946459966905,0.050129387245278045,0.05533147566918936,0.06395658952859787,1.0983753514559728e-19,0.07725062357811513,0.0839247261204711,0.019340663154145395,72.75871981447884,1.9983123784016887,0.0707878956255097,0.02412324759507635,0.07930911307602667,0.016998249475751393,0.009591669882085423,0.016998249475751393,0.009591669882085426,0.2942328226150227,0.2892696425372333,0.2872103195126915,1.7574005623295565e-18,3.514801124659113e-18
6
+ min,0.0,0.0,0.0,0.0,0.0,0.03144016227180527,0.0,0.01,0.01,0.034,-0.0009229119228736442,0.6057692307692307,0.49227568547956896,0.9074074074074074,32.0,983.0,0.03383549493412961,0.44615314358781144,-0.14814114995900682,0.114609793407313,0.15071213597912292,0.114609793407313,0.15071213597912286,0.02564102564102566,0.02564102564102566,0.02564102564102566,0.011980186563896641,0.02388874024156554
7
+ 25%,0.04580152671755725,0.03125,0.05,0.04580152671755725,0.04580152671755725,0.09615384615384616,0.04,0.05,0.04,0.054,-0.0009229119228736442,0.7094017094017094,0.6864788579493025,0.9357798165137615,103.0,986.0,0.1595478769023866,0.5001644156427448,-0.04447327870140726,0.1347149541668744,0.16294666389985077,0.1347149541668744,0.16294666389985077,0.33333333333333337,0.3076923076923077,0.3076923076923077,0.011980186563896641,0.02388874024156554
8
+ 50%,0.06164383561643835,0.06,0.06,0.06164383561643835,0.06164383561643835,0.11212121212121212,0.06,0.06,0.055,0.072,-0.0009229119228736442,0.7580645161290323,0.7473073968239867,0.9473684210526315,117.0,987.0,0.1987413484417469,0.5102200996677742,-0.009140626573123105,0.14049660342548265,0.16901216639230882,0.14049660342548265,0.16901216639230882,0.5897435897435898,0.5128205128205128,0.5384615384615384,0.011980186563896641,0.02388874024156554
9
+ 75%,0.07633587786259542,0.08,0.07,0.07633587786259542,0.07633587786259542,0.15384615384615385,0.08,0.07,0.08,0.1,-0.0009229119228736442,0.8130841121495327,0.7853363113871351,0.9629629629629629,158.0,989.0,0.23113725896307896,0.5261389096340552,0.04111513047747314,0.15225451226071351,0.1734566066129824,0.15225451226071351,0.17345660661298243,0.717948717948718,0.7692307692307692,0.7435897435897436,0.011980186563896641,0.02388874024156554
10
+ max,0.28720626631853785,0.18,0.24,0.27,0.28720626631853785,0.3785425101214575,0.18,0.24,0.27,0.314,-0.0009229119228736442,0.8901960784313725,0.8533279423212086,0.9770992366412213,384.0,992.0,0.40325964646448925,0.5509811721028905,0.27641054721277936,0.2007382929173619,0.200196903370157,0.2007382929173619,0.200196903370157,1.0,1.0,1.0,0.011980186563896641,0.02388874024156554
transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89772db631dc8ae3640e65a6a0f908f7ddb80bb13982d884d455e79b67507983
3
+ size 61460400
transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/pred.h5ad ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c5f24c6512fad9bcc7c8d3e9b26fb5a78906d094f00a99cf231b3e524c3e563
3
+ size 49952825
transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/real.h5ad ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d906eb6262ae9d587a091c8f6722e562182c9cd62e18f7c2faecbd7c4935ec3c
3
+ size 75069953
transfer/code/CCFM/result/ccfm-fusion_differential_perceiver-norman-cascaded-predict_y-gamma_0.5-perturbation_function_crisper-lr_5e-05-dim_model_128-infer_top_gene_1000-split_method_additive-use_mmd_loss_True-fold_1-latent_weight_1.0-choose_latent_p_0.4/iteration_110000/results.csv ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ perturbation,overlap_at_N,overlap_at_50,overlap_at_100,overlap_at_200,overlap_at_500,precision_at_N,precision_at_50,precision_at_100,precision_at_200,precision_at_500,de_spearman_sig,de_direction_match,de_spearman_lfc_sig,de_sig_genes_recall,de_nsig_counts_real,de_nsig_counts_pred,pr_auc,roc_auc,pearson_delta,mse,mae,mse_delta,mae_delta,discrimination_score_l1,discrimination_score_l2,discrimination_score_cosine,pearson_edistance,clustering_agreement
2
+ AHR+FEV,0.11682242990654206,0.14,0.15,0.12,0.11682242990654206,0.20466058763931105,0.14,0.15,0.12,0.154,-0.0009229119228736442,0.8130841121495327,0.7853363113871351,0.9439252336448598,214.0,987.0,0.2526840907570853,0.5084094314047228,0.11135915537214161,0.13533302491436547,0.1613838711025711,0.13533302491436547,0.1613838711025711,0.17948717948717952,0.3589743589743589,0.9743589743589743,0.011980186563896641,0.02388874024156554
3
+ AHR+KLF1,0.14285714285714285,0.18,0.17,0.14285714285714285,0.14285714285714285,0.13333333333333333,0.18,0.17,0.12,0.092,-0.0009229119228736442,0.7857142857142857,0.7671530770745367,0.9428571428571428,140.0,990.0,0.20910017979810214,0.5102200996677742,-0.12253033415746277,0.13758035051028378,0.15343646898440932,0.13758035051028378,0.15343646898440932,0.28205128205128205,0.3589743589743589,0.07692307692307687,0.011980186563896641,0.02388874024156554
4
+ BCL2L11+BAK1,0.03125,0.03125,0.03125,0.03125,0.03125,0.03144016227180527,0.04,0.04,0.045,0.036,-0.0009229119228736442,0.84375,0.7455545371219066,0.96875,32.0,986.0,0.03383549493412961,0.4650051652892562,0.0735293611513918,0.11762834218056595,0.16137206790715886,0.11762834218056595,0.16137206790715886,0.717948717948718,0.9230769230769231,0.7435897435897436,0.011980186563896641,0.02388874024156554
5
+ BPGM+ZBTB1,0.045454545454545456,0.04,0.05,0.045454545454545456,0.045454545454545456,0.1033434650455927,0.04,0.05,0.06,0.062,-0.0009229119228736442,0.7181818181818181,0.7059420361744055,0.9272727272727272,110.0,987.0,0.126474476292699,0.46615934627170585,-0.03340010334503688,0.1470699804162752,0.17257140861649084,0.1470699804162752,0.17257140861649084,0.46153846153846156,0.4358974358974359,0.33333333333333337,0.011980186563896641,0.02388874024156554
6
+ CBL+PTPN12,0.04807692307692308,0.08,0.05,0.04807692307692308,0.04807692307692308,0.09615384615384616,0.08,0.05,0.04,0.054,-0.0009229119228736442,0.6057692307692307,0.49227568547956896,0.9134615384615384,104.0,988.0,0.15494497852716124,0.47358988667582413,-0.04708901684337179,0.1412088514969795,0.15595579153966155,0.1412088514969795,0.15595579153966155,0.33333333333333337,0.23076923076923073,0.17948717948717952,0.011980186563896641,0.02388874024156554
7
+ CBL+PTPN9,0.05263157894736842,0.04,0.05263157894736842,0.05263157894736842,0.05263157894736842,0.09008097165991903,0.04,0.05,0.04,0.05,-0.0009229119228736442,0.6631578947368421,0.5791315546347882,0.9368421052631579,95.0,988.0,0.22980702723803084,0.5229659784821169,-0.009140626573123105,0.135391186039481,0.16404357622867724,0.135391186039481,0.16404357622867724,0.5897435897435898,0.5641025641025641,0.3846153846153846,0.011980186563896641,0.02388874024156554
8
+ CBL+TGFBR2,0.0136986301369863,0.02,0.0136986301369863,0.0136986301369863,0.0136986301369863,0.06983805668016195,0.02,0.02,0.045,0.048,-0.0009229119228736442,0.6575342465753424,0.6983955569268744,0.9452054794520548,73.0,988.0,0.14938762945256867,0.5269391615315276,0.011487003874861545,0.1347149541668744,0.16475300432481219,0.1347149541668744,0.16475300432481219,0.6666666666666667,0.6923076923076923,0.5384615384615384,0.011980186563896641,0.02388874024156554
9
+ CBL+UBASH3A,0.0,0.0,0.0,0.0,0.0,0.044444444444444446,0.0,0.01,0.01,0.038,-0.0009229119228736442,0.7872340425531915,0.8304810360777058,0.9361702127659575,47.0,990.0,0.08497273037937422,0.5139871849255432,-0.03165417907154238,0.14438255708270986,0.17027819520042398,0.14438255708270986,0.17027819520042398,0.6666666666666667,0.5641025641025641,0.3589743589743589,0.011980186563896641,0.02388874024156554
10
+ CBL+UBASH3B,0.06164383561643835,0.08,0.07,0.06164383561643835,0.06164383561643835,0.13793103448275862,0.08,0.07,0.07,0.074,-0.0009229119228736442,0.636986301369863,0.6095019393682098,0.9315068493150684,146.0,986.0,0.25316040825284114,0.5001644156427448,0.014004357377824761,0.13739740766424166,0.15257307516449997,0.13739740766424166,0.15257307516449997,0.3846153846153846,0.2564102564102564,0.5641025641025641,0.011980186563896641,0.02388874024156554
11
+ CDKN1B+CDKN1A,0.04854368932038835,0.08,0.05,0.04854368932038835,0.04854368932038835,0.09929078014184398,0.08,0.05,0.045,0.074,-0.0009229119228736442,0.7864077669902912,0.8389491719017704,0.9514563106796117,103.0,987.0,0.17336435770437908,0.5261389096340552,-0.014567903048687377,0.13772874508480135,0.1725944370902679,0.13772874508480135,0.17259443709026787,0.6666666666666667,0.6666666666666667,0.5128205128205128,0.011980186563896641,0.02388874024156554
12
+ CDKN1C+CDKN1B,0.06363636363636363,0.0,0.07,0.06363636363636363,0.06363636363636363,0.1082995951417004,0.0,0.07,0.05,0.086,-0.0009229119228736442,0.8,0.8120677684907175,0.9727272727272728,110.0,988.0,0.1595478769023866,0.5272727272727274,-0.015993003377827556,0.14712793306264882,0.17606891978883893,0.14712793306264882,0.17606891978883893,0.6923076923076923,0.717948717948718,0.5384615384615384,0.011980186563896641,0.02388874024156554
13
+ CEBPE+CEBPA,0.28720626631853785,0.14,0.24,0.255,0.28720626631853785,0.3746192893401015,0.14,0.24,0.255,0.314,-0.0009229119228736442,0.8825065274151436,0.7497669895999942,0.9634464751958225,383.0,985.0,0.40162534646067033,0.5036456195437368,0.053290754365770444,0.2007382929173619,0.200196903370157,0.2007382929173619,0.200196903370157,0.02564102564102566,0.02564102564102566,0.9230769230769231,0.011980186563896641,0.02388874024156554
14
+ CEBPE+CNN1,0.09523809523809523,0.02,0.07,0.09523809523809523,0.09523809523809523,0.1431472081218274,0.02,0.07,0.11,0.112,-0.0009229119228736442,0.8639455782312925,0.8533279423212086,0.9591836734693877,147.0,985.0,0.23359164712607777,0.5485481414136579,-0.005738402924381779,0.13665777832564358,0.16651085637358815,0.13665777832564358,0.16651085637358815,1.0,0.7435897435897436,0.6153846153846154,0.011980186563896641,0.02388874024156554
15
+ ETS2+IGDCC3,0.10434782608695652,0.08,0.11,0.10434782608695652,0.10434782608695652,0.11032388663967611,0.08,0.11,0.08,0.07,-0.0009229119228736442,0.7565217391304347,0.7767349709522606,0.9478260869565217,115.0,988.0,0.2299799878835566,0.5391697371653156,0.0669259739814912,0.12840725080454235,0.16921041898532832,0.12840725080454235,0.16921041898532832,0.8205128205128205,0.8974358974358975,0.8717948717948718,0.011980186563896641,0.02388874024156554
16
+ ETS2+IKZF3,0.07526881720430108,0.08,0.07,0.07526881720430108,0.07526881720430108,0.18163471241170534,0.08,0.07,0.08,0.1,-0.0009229119228736442,0.7580645161290323,0.6771519332764242,0.967741935483871,186.0,991.0,0.260131332480487,0.5207028876383716,-0.09463513777294134,0.17327119781195713,0.17342775298852126,0.17327119781195713,0.17342775298852128,0.1282051282051282,0.07692307692307687,0.1282051282051282,0.011980186563896641,0.02388874024156554
17
+ FEV+ISL2,0.05921052631578947,0.04,0.05,0.05921052631578947,0.05921052631578947,0.1440162271805274,0.04,0.05,0.065,0.084,-0.0009229119228736442,0.7894736842105263,0.7834561274551701,0.9342105263157895,152.0,986.0,0.20090127062097754,0.4942317837636544,-0.04447327870140726,0.14407596376588572,0.17563518809557005,0.14407596376588572,0.17563518809557005,0.23076923076923073,0.3076923076923077,0.3589743589743589,0.011980186563896641,0.02388874024156554
18
+ FOSB+CEBPB,0.19607843137254902,0.08,0.13,0.175,0.19607843137254902,0.24974721941354905,0.08,0.13,0.175,0.228,-0.0009229119228736442,0.8901960784313725,0.7759819605912757,0.9686274509803922,255.0,989.0,0.2863141384524899,0.5037557573364916,0.04111513047747314,0.1649500859024486,0.18023598317837822,0.1649500859024486,0.18023598317837822,0.05128205128205132,0.05128205128205132,0.8461538461538461,0.011980186563896641,0.02388874024156554
19
+ FOXA3+FOXA1,0.07142857142857142,0.14,0.08,0.07142857142857142,0.07142857142857142,0.1337386018237082,0.14,0.08,0.08,0.104,-0.0009229119228736442,0.7857142857142857,0.8403498797288432,0.9428571428571428,140.0,987.0,0.19499722204763556,0.4847259136212623,0.028711468025650683,0.12510979604908035,0.16607759556961435,0.12510979604908035,0.16607759556961435,0.9230769230769231,0.7948717948717949,0.7435897435897436,0.011980186563896641,0.02388874024156554
20
+ KLF1+BAK1,0.0625,0.06,0.0625,0.0625,0.0625,0.09155645981688708,0.06,0.06,0.05,0.062,-0.0009229119228736442,0.6354166666666666,0.6864788579493025,0.9375,96.0,983.0,0.18343355243065262,0.5053350848082596,-0.13773340365456174,0.1368285814597498,0.16320140548731554,0.1368285814597498,0.16320140548731554,0.20512820512820518,0.4871794871794872,0.02564102564102566,0.011980186563896641,0.02388874024156554
21
+ KLF1+CEBPA,0.2734375,0.12,0.2,0.27,0.2734375,0.3785425101214575,0.12,0.2,0.27,0.31,-0.0009229119228736442,0.8828125,0.7503365950725829,0.9739583333333334,384.0,988.0,0.40325964646448925,0.5037899925595237,0.053321696247698225,0.1635201302608695,0.18066548643222594,0.1635201302608695,0.1806654864322259,0.07692307692307687,0.1282051282051282,0.8974358974358975,0.011980186563896641,0.02388874024156554
22
+ KLF1+CLDN6,0.07633587786259542,0.04,0.07,0.07633587786259542,0.07633587786259542,0.12195121951219512,0.04,0.07,0.075,0.092,-0.0009229119228736442,0.7022900763358778,0.7527439675421739,0.916030534351145,131.0,984.0,0.20040341203974787,0.5151925087184532,-0.14814114995900682,0.1749716160742202,0.18067823362336982,0.1749716160742202,0.18067823362336982,0.10256410256410253,0.1282051282051282,0.05128205128205132,0.011980186563896641,0.02388874024156554
23
+ LYL1+CEBPB,0.06962025316455696,0.04,0.06,0.06962025316455696,0.06962025316455696,0.15384615384615385,0.04,0.06,0.07,0.112,-0.0009229119228736442,0.8227848101265823,0.8350167530440022,0.9620253164556962,158.0,988.0,0.21810176703580236,0.5133572867494512,-0.10050849461118233,0.15142331243063206,0.17883328209911836,0.15142331243063206,0.17883328209911836,0.6666666666666667,0.5641025641025641,0.15384615384615385,0.011980186563896641,0.02388874024156554
24
+ MAP2K3+ELMSAN1,0.03896103896103896,0.04,0.03,0.03896103896103896,0.03896103896103896,0.14705882352941177,0.04,0.03,0.045,0.084,-0.0009229119228736442,0.7857142857142857,0.7473073968239867,0.9415584415584416,154.0,986.0,0.1987413484417469,0.5016195388535815,0.04138578683734644,0.12619019587427552,0.16987441427815242,0.12619019587427552,0.16987441427815242,0.8461538461538461,0.8461538461538461,0.717948717948718,0.011980186563896641,0.02388874024156554
25
+ MAP2K3+IKZF3,0.046296296296296294,0.06,0.05,0.046296296296296294,0.046296296296296294,0.09939148073022312,0.06,0.05,0.05,0.054,-0.0009229119228736442,0.7222222222222222,0.699514808354066,0.9074074074074074,108.0,986.0,0.16947535116567436,0.5236308337485468,-0.03661169397625287,0.15362600905371226,0.16178276907210315,0.15362600905371226,0.16178276907210315,0.10256410256410253,0.07692307692307687,0.3076923076923077,0.011980186563896641,0.02388874024156554
26
+ MAP2K3+MAP2K6,0.028169014084507043,0.02,0.028169014084507043,0.028169014084507043,0.028169014084507043,0.06896551724137931,0.02,0.03,0.025,0.034,-0.0009229119228736442,0.6619718309859155,0.6651278768069548,0.9577464788732394,71.0,986.0,0.11733973805361825,0.5203535529647204,0.10855741581134039,0.114609793407313,0.15071213597912292,0.114609793407313,0.15071213597912286,0.8461538461538461,0.9743589743589743,0.9230769230769231,0.011980186563896641,0.02388874024156554
27
+ MAP2K3+SLC38A2,0.0,0.0,0.0,0.0,0.0,0.054600606673407485,0.0,0.03,0.02,0.034,-0.0009229119228736442,0.7719298245614035,0.706948051948052,0.9473684210526315,57.0,989.0,0.12075484905850475,0.48587003032501713,0.27641054721277936,0.11832837886785864,0.15628214561157314,0.11832837886785864,0.15628214561157314,1.0,1.0,1.0,0.011980186563896641,0.02388874024156554
28
+ MAP2K6+ELMSAN1,0.06976744186046512,0.1,0.06976744186046512,0.06976744186046512,0.06976744186046512,0.08121827411167512,0.1,0.07,0.055,0.054,-0.0009229119228736442,0.8255813953488372,0.7978648089938549,0.9302325581395349,86.0,985.0,0.1488116939492045,0.5116597119739453,0.08235530360106362,0.13344647098238938,0.17605625595881952,0.13344647098238938,0.17605625595881952,0.8205128205128205,0.9487179487179487,0.8717948717948718,0.011980186563896641,0.02388874024156554
29
+ MAPK1+IKZF3,0.06707317073170732,0.06,0.04,0.06707317073170732,0.06707317073170732,0.15821501014198783,0.06,0.04,0.07,0.1,-0.0009229119228736442,0.7378048780487805,0.7133785429094198,0.9512195121951219,164.0,986.0,0.24844090370868255,0.5299043062200957,-0.1047008090643299,0.1439465863109621,0.16327971014530332,0.1439465863109621,0.16327971014530332,0.46153846153846156,0.23076923076923073,0.10256410256410253,0.011980186563896641,0.02388874024156554
30
+ PTPN12+PTPN9,0.04081632653061224,0.06,0.04081632653061224,0.04081632653061224,0.04081632653061224,0.09017223910840932,0.06,0.05,0.025,0.046,-0.0009229119228736442,0.673469387755102,0.5790681954417335,0.9081632653061225,98.0,987.0,0.14723213237377317,0.4687372731797819,-0.05982258053324594,0.13596932558839342,0.1676436597212595,0.13596932558839342,0.16764365972125952,0.5897435897435898,0.46153846153846156,0.2564102564102564,0.011980186563896641,0.02388874024156554
31
+ PTPN12+UBASH3A,0.05217391304347826,0.02,0.05,0.05217391304347826,0.05217391304347826,0.10797174571140263,0.02,0.05,0.055,0.064,-0.0009229119228736442,0.6521739130434783,0.5982158888898341,0.9304347826086956,115.0,991.0,0.18825985678860804,0.4751166789486612,-0.04315213343051622,0.15225451226071351,0.16901216639230882,0.15225451226071351,0.16901216639230882,0.3589743589743589,0.3076923076923077,0.28205128205128205,0.011980186563896641,0.02388874024156554
32
+ SGK1+S1PR2,0.09230769230769231,0.16,0.16,0.09230769230769231,0.09230769230769231,0.1896551724137931,0.16,0.16,0.09,0.106,-0.0009229119228736442,0.8051282051282052,0.7739305882677583,0.958974358974359,195.0,986.0,0.23113725896307896,0.5063991081382385,0.004528795245680833,0.15373867956576176,0.1734566066129824,0.15373867956576176,0.17345660661298243,0.6666666666666667,0.5384615384615384,0.6923076923076923,0.011980186563896641,0.02388874024156554
33
+ SGK1+TBX2,0.08196721311475409,0.12,0.09,0.08196721311475409,0.08196721311475409,0.1172901921132457,0.12,0.09,0.065,0.074,-0.0009229119228736442,0.7540983606557377,0.731259317042213,0.9508196721311475,122.0,989.0,0.1809233305881876,0.5103532618843124,-0.06813754787056124,0.14238519904308378,0.16172195157060876,0.14238519904308375,0.16172195157060876,0.5897435897435898,0.46153846153846156,0.3076923076923077,0.011980186563896641,0.02388874024156554
34
+ SGK1+TBX3,0.045871559633027525,0.02,0.05,0.045871559633027525,0.045871559633027525,0.10707070707070707,0.02,0.05,0.04,0.062,-0.0009229119228736442,0.7339449541284404,0.6940824233885171,0.9724770642201835,109.0,990.0,0.22068873421549678,0.5324859193360724,-0.06788124274572588,0.15605050105133378,0.17120755159227555,0.15605050105133378,0.17120755159227555,0.3589743589743589,0.2564102564102564,0.33333333333333337,0.011980186563896641,0.02388874024156554
35
+ TBX3+TBX2,0.030864197530864196,0.06,0.04,0.030864197530864196,0.030864197530864196,0.1592741935483871,0.06,0.04,0.025,0.052,-0.0009229119228736442,0.7283950617283951,0.6606624947093028,0.9753086419753086,162.0,992.0,0.2804595874205586,0.5509811721028905,-0.003876782304558205,0.1552928291345333,0.17319104731062776,0.1552928291345333,0.17319104731062776,0.4358974358974359,0.4871794871794872,0.5641025641025641,0.011980186563896641,0.02388874024156554
36
+ TGFBR2+IGDCC3,0.05504587155963303,0.02,0.05,0.05504587155963303,0.05504587155963303,0.1033434650455927,0.02,0.05,0.04,0.052,-0.0009229119228736442,0.6422018348623854,0.6404013253933313,0.9357798165137615,109.0,987.0,0.1836325432293462,0.49854302453690835,0.029582334662403266,0.13419100933442726,0.17173901466956068,0.13419100933442726,0.17173901466956068,0.5384615384615384,0.7692307692307692,0.6666666666666667,0.011980186563896641,0.02388874024156554
37
+ TGFBR2+PRTG,0.06837606837606838,0.1,0.06,0.06837606837606838,0.06837606837606838,0.11212121212121212,0.1,0.06,0.05,0.07,-0.0009229119228736442,0.7094017094017094,0.7512478453121487,0.9487179487179487,117.0,990.0,0.17398226333930353,0.5051059422520351,0.0051239719507837666,0.15126123882176176,0.18133189622175813,0.15126123882176176,0.18133189622175816,0.41025641025641024,0.6666666666666667,0.6153846153846154,0.011980186563896641,0.02388874024156554
38
+ UBASH3B+OSR2,0.04580152671755725,0.04,0.06,0.04580152671755725,0.04580152671755725,0.12942366026289182,0.04,0.06,0.03,0.072,-0.0009229119228736442,0.8473282442748091,0.8039365452408931,0.9770992366412213,131.0,989.0,0.20597696772069213,0.534254517344671,-0.024733824741466907,0.1269579840157475,0.16294666389985077,0.1269579840157475,0.16294666389985077,0.9743589743589743,0.8205128205128205,0.5128205128205128,0.011980186563896641,0.02388874024156554
39
+ UBASH3B+ZBTB25,0.030927835051546393,0.02,0.030927835051546393,0.030927835051546393,0.030927835051546393,0.09008097165991903,0.02,0.03,0.03,0.044,-0.0009229119228736442,0.711340206185567,0.6293461176764588,0.9175257731958762,97.0,988.0,0.1417607979403474,0.44615314358781144,-0.02539219177148751,0.14049660342548265,0.16065800141924427,0.14049660342548265,0.16065800141924427,0.6666666666666667,0.5128205128205128,0.4358974358974359,0.011980186563896641,0.02388874024156554
40
+ ZC3HAV1+CEBPE,0.07407407407407407,0.06,0.06,0.07407407407407407,0.07407407407407407,0.1580547112462006,0.06,0.06,0.08,0.122,-0.0009229119228736442,0.8333333333333334,0.7982736441978715,0.9629629629629629,162.0,987.0,0.2208512850815783,0.5424106485164559,0.03838944323594112,0.12623136181133277,0.1662241610182183,0.12623136181133277,0.1662241610182183,0.9743589743589743,0.8717948717948718,0.7692307692307692,0.011980186563896641,0.02388874024156554
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/iteration_10000/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eef965a2864046958117a5b478bd5b23ec82e6878b9ff6402ed9c8b8741794f7
3
+ size 61460656
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/loss_curve.csv ADDED
The diff for this file is too large to render. See raw diff
 
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/tb_logs/events.out.tfevents.1773518027.b0021.201.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f71267f8ecf067010bbdb97d46a2bfcca9d7c1a64c668ccae8b61ed18ab483d
3
+ size 88
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/tb_logs/events.out.tfevents.1773518372.b0021.200.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35c7183043560c21f8f4392ebed984d4dd7c7b6df04b5602299592a177ccc770
3
+ size 88
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/tb_logs/events.out.tfevents.1773518599.b0021.200.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f816e022574e0545cf550251c915e782d6c62a2fc6733f23dccca1885fd6efa
3
+ size 88
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-cached/tb_logs/events.out.tfevents.1773561648.b0034.161.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9f84c06bf4e08b674f8b5008d1edbada1c1c6d31f3959496d7badc2e35305ca
3
+ size 1091934
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/eval_only/agg_results.csv ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ statistic,overlap_at_N,overlap_at_50,overlap_at_100,overlap_at_200,overlap_at_500,precision_at_N,precision_at_50,precision_at_100,precision_at_200,precision_at_500,de_spearman_sig,de_direction_match,de_spearman_lfc_sig,de_sig_genes_recall,de_nsig_counts_real,de_nsig_counts_pred,pr_auc,roc_auc,pearson_delta,mse,mae,mse_delta,mae_delta,discrimination_score_l1,discrimination_score_l2,discrimination_score_cosine,pearson_edistance,clustering_agreement
2
+ count,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0
3
+ null_count,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
4
+ mean,0.07434137778088827,0.07157051282051285,0.07604612557634062,0.07496692561581857,0.07434137778088827,0.13088753921097923,0.07179487179487183,0.07666666666666669,0.07435897435897437,0.06779487179487183,-0.20008264511131116,0.7294312841974797,0.7296732981115676,0.9273631404325153,137.89743589743588,982.5641025641025,0.23830955747498037,0.5247149499857076,-0.11515337322550942,0.04572970260837487,0.06046652511863345,0.04572970260837487,0.06046652511863345,0.517422748191979,0.5154503616042078,0.5187376725838265,-0.03164181640522182,0.09980551725234033
5
+ std,0.06538979015358905,0.04845110627973415,0.06659633696295453,0.06675561380795758,0.06538979015358905,0.07185080468596214,0.048279438517196564,0.06618686995459118,0.06561630380657058,0.06212007874701836,2.8118408997272905e-17,0.05319787178851119,0.08148598126037082,0.02178139403957795,72.75871981447884,2.173947074779103,0.0707575686657406,0.024027439137493466,0.13199828903665942,0.01861458373584706,0.013575957811586133,0.018614583735847057,0.013575957811586135,0.29329047295895705,0.2910355912021317,0.2889274413247139,1.4059204498636452e-17,2.8118408997272905e-17
6
+ min,0.0,0.0,0.0,0.0,0.0,0.03147208121827411,0.0,0.0,0.005,0.014,-0.2000826451113112,0.6330275229357798,0.5557406422352249,0.8645833333333334,32.0,976.0,0.0516144925026014,0.4870105625153525,-0.31587954989032513,0.02171010429886536,0.0479999642064419,0.02171010429886536,0.0479999642064419,0.02564102564102566,0.02564102564102566,0.02564102564102566,-0.031641816405221804,0.09980551725234024
7
+ 25%,0.04081632653061224,0.04,0.04,0.04081632653061224,0.04081632653061224,0.09693877551020408,0.04,0.04,0.045,0.04,-0.2000826451113112,0.6870229007633588,0.6792914830965147,0.9178082191780822,103.0,981.0,0.19439001023632266,0.5110010214504597,-0.19877860618913304,0.0360505190620416,0.054246974930411994,0.0360505190620416,0.054246974930411994,0.28205128205128205,0.3076923076923077,0.28205128205128205,-0.031641816405221804,0.09980551725234024
8
+ 50%,0.04854368932038835,0.06,0.05,0.04854368932038835,0.04854368932038835,0.10986775178026449,0.06,0.05,0.055,0.05,-0.2000826451113112,0.7211538461538461,0.7467580642404144,0.9230769230769231,117.0,983.0,0.23279528674164413,0.5186901913875597,-0.1437614702536516,0.04238467894650805,0.056591795566055876,0.04238467894650805,0.056591795566055876,0.4871794871794872,0.5128205128205128,0.5384615384615384,-0.031641816405221804,0.09980551725234024
9
+ 75%,0.08860759493670886,0.1,0.1,0.08860759493670886,0.08860759493670886,0.15061475409836064,0.1,0.1,0.08,0.072,-0.2000826451113112,0.75,0.7934782608695652,0.9389312977099237,158.0,984.0,0.2765077761090199,0.5394629900548241,-0.03567863213268257,0.05394683661706369,0.060919073434818234,0.05394683661706369,0.060919073434818234,0.7692307692307692,0.7435897435897436,0.7692307692307692,-0.031641816405221804,0.09980551725234024
10
+ max,0.3203125,0.2,0.31,0.325,0.3203125,0.37741607324516785,0.2,0.31,0.325,0.318,-0.2000826451113112,0.8567708333333334,0.868395024908931,0.9753086419753086,384.0,986.0,0.43969830739977106,0.5976688392590536,0.30981581535034897,0.11725468655875379,0.11637918515028178,0.11725468655875378,0.1163791851502818,1.0,1.0,1.0,-0.031641816405221804,0.09980551725234024
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/eval_only/pred.h5ad ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f38618e6770118329326dabc7a9c8d99534bdd45eb0aa9f2887144d3292ebcda
3
+ size 49952825
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/eval_only/real.h5ad ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d906eb6262ae9d587a091c8f6722e562182c9cd62e18f7c2faecbd7c4935ec3c
3
+ size 75069953
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/eval_only/results.csv ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ perturbation,overlap_at_N,overlap_at_50,overlap_at_100,overlap_at_200,overlap_at_500,precision_at_N,precision_at_50,precision_at_100,precision_at_200,precision_at_500,de_spearman_sig,de_direction_match,de_spearman_lfc_sig,de_sig_genes_recall,de_nsig_counts_real,de_nsig_counts_pred,pr_auc,roc_auc,pearson_delta,mse,mae,mse_delta,mae_delta,discrimination_score_l1,discrimination_score_l2,discrimination_score_cosine,pearson_edistance,clustering_agreement
2
+ AHR+FEV,0.14485981308411214,0.18,0.14,0.145,0.14485981308411214,0.20710659898477157,0.18,0.14,0.145,0.104,-0.2000826451113112,0.7897196261682243,0.8073102807791784,0.9532710280373832,214.0,985.0,0.29945254193043386,0.512624551140282,-0.03661475927802558,0.06221497299106943,0.07498016592692698,0.06221497299106943,0.07498016592692698,0.10256410256410253,0.15384615384615385,0.717948717948718,-0.031641816405221804,0.09980551725234024
3
+ AHR+KLF1,0.10714285714285714,0.16,0.1,0.10714285714285714,0.10714285714285714,0.13109756097560976,0.16,0.1,0.095,0.068,-0.2000826451113112,0.6714285714285714,0.760594193703647,0.9214285714285714,140.0,984.0,0.22043283181034135,0.4962209302325581,-0.29936661680170734,0.04578033190197752,0.059630842494927554,0.04578033190197752,0.059630842494927554,0.3076923076923077,0.41025641025641024,0.07692307692307687,-0.031641816405221804,0.09980551725234024
4
+ BCL2L11+BAK1,0.03125,0.03125,0.03125,0.03125,0.03125,0.03147208121827411,0.04,0.04,0.045,0.034,-0.2000826451113112,0.8125,0.8397800183318057,0.96875,32.0,985.0,0.0516144925026014,0.5442600723140496,0.00046406370155914435,0.025391938886468096,0.0479999642064419,0.025391938886468096,0.0479999642064419,0.9487179487179487,0.9487179487179487,0.8461538461538461,-0.031641816405221804,0.09980551725234024
5
+ BPGM+ZBTB1,0.05454545454545454,0.02,0.06,0.05454545454545454,0.05454545454545454,0.10408163265306122,0.02,0.06,0.05,0.046,-0.2000826451113112,0.7181818181818181,0.6792914830965147,0.9272727272727272,110.0,980.0,0.17902130967342683,0.5110010214504597,-0.23073459407314534,0.04550141186047617,0.055509060403470895,0.04550141186047617,0.055509060403470895,0.4871794871794872,0.3846153846153846,0.20512820512820518,-0.031641816405221804,0.09980551725234024
6
+ CBL+PTPN12,0.04807692307692308,0.08,0.05,0.04807692307692308,0.04807692307692308,0.0973630831643002,0.08,0.05,0.045,0.03,-0.2000826451113112,0.7211538461538461,0.5591699562573349,0.9230769230769231,104.0,986.0,0.26403848667904933,0.5294041895604397,-0.23274889893333786,0.05862249325747039,0.06036412591726626,0.05862249325747039,0.06036412591726626,0.41025641025641024,0.2564102564102564,0.20512820512820518,-0.031641816405221804,0.09980551725234024
7
+ CBL+PTPN9,0.042105263157894736,0.06,0.042105263157894736,0.042105263157894736,0.042105263157894736,0.08952187182095625,0.06,0.04,0.04,0.03,-0.2000826451113112,0.7684210526315789,0.5557406422352249,0.9263157894736842,95.0,983.0,0.24315418177416312,0.5552311718522827,-0.1437614702536516,0.03940013269022343,0.053452194004056305,0.03940013269022343,0.053452194004056305,0.7692307692307692,0.5897435897435898,0.5897435897435898,-0.031641816405221804,0.09980551725234024
8
+ CBL+TGFBR2,0.0410958904109589,0.06,0.0410958904109589,0.0410958904109589,0.0410958904109589,0.0681586978636826,0.06,0.05,0.06,0.036,-0.2000826451113112,0.684931506849315,0.6195618636223388,0.9178082191780822,73.0,983.0,0.1655895799050663,0.5394629900548241,-0.20560209563646858,0.04304595012673362,0.054979362744311845,0.04304595012673362,0.054979362744311845,0.7435897435897436,0.6666666666666667,0.4358974358974359,-0.031641816405221804,0.09980551725234024
9
+ CBL+UBASH3A,0.0,0.0,0.0,0.0,0.0,0.04369918699186992,0.0,0.0,0.005,0.014,-0.2000826451113112,0.7021276595744681,0.7934782608695652,0.9148936170212766,47.0,984.0,0.14696506452755298,0.5480118773860821,-0.18161296667535468,0.04238467894650805,0.05764305930724554,0.04238467894650805,0.05764305930724554,0.5641025641025641,0.4871794871794872,0.28205128205128205,-0.031641816405221804,0.09980551725234024
10
+ CBL+UBASH3B,0.0821917808219178,0.14,0.08,0.0821917808219178,0.0821917808219178,0.13807106598984772,0.14,0.08,0.06,0.05,-0.2000826451113112,0.7465753424657534,0.6421219196850697,0.9315068493150684,146.0,985.0,0.29997837940231675,0.5321532835006898,-0.15786369842915818,0.05987722901827767,0.05868575068700038,0.05987722901827767,0.05868575068700038,0.46153846153846156,0.20512820512820518,0.4871794871794872,-0.031641816405221804,0.09980551725234024
11
+ CDKN1B+CDKN1A,0.04854368932038835,0.04,0.05,0.04854368932038835,0.04854368932038835,0.09693877551020408,0.04,0.05,0.055,0.05,-0.2000826451113112,0.7184466019417476,0.8072090673461318,0.9223300970873787,103.0,980.0,0.1694843551253765,0.5162082886861273,-0.19937791431428054,0.037691983646484795,0.057289317060585716,0.037691983646484795,0.057289317060585716,0.4358974358974359,0.6153846153846154,0.33333333333333337,-0.031641816405221804,0.09980551725234024
12
+ CDKN1C+CDKN1B,0.05454545454545454,0.06,0.06,0.05454545454545454,0.05454545454545454,0.10488798370672098,0.06,0.06,0.05,0.046,-0.2000826451113112,0.7454545454545455,0.8149079859702273,0.9363636363636364,110.0,982.0,0.18638794131237005,0.5296373850868232,-0.16090030330901103,0.03772396068334866,0.055244365061252895,0.03772396068334866,0.055244365061252895,0.641025641025641,0.6923076923076923,0.3589743589743589,-0.031641816405221804,0.09980551725234024
13
+ CEBPE+CEBPA,0.2950391644908616,0.06,0.3,0.325,0.2950391644908616,0.3733468972533062,0.06,0.3,0.325,0.294,-0.2000826451113112,0.8511749347258486,0.7521492072928287,0.95822454308094,383.0,983.0,0.4115193409218539,0.5048051085222439,0.02015192795065466,0.11725468655875379,0.11637918515028178,0.11725468655875378,0.1163791851502818,0.02564102564102566,0.02564102564102566,0.8717948717948718,-0.031641816405221804,0.09980551725234024
14
+ CEBPE+CNN1,0.11564625850340136,0.14,0.11,0.11564625850340136,0.11564625850340136,0.1396534148827727,0.14,0.11,0.095,0.088,-0.2000826451113112,0.7414965986394558,0.8447312933138854,0.9319727891156463,147.0,981.0,0.2765077761090199,0.5520332400252012,-0.05815205060952569,0.03218865293720686,0.056591795566055876,0.03218865293720686,0.056591795566055876,0.4871794871794872,0.7948717948717949,0.7435897435897436,-0.031641816405221804,0.09980551725234024
15
+ ETS2+IGDCC3,0.06956521739130435,0.08,0.08,0.06956521739130435,0.06956521739130435,0.109072375127421,0.08,0.08,0.075,0.06,-0.2000826451113112,0.6869565217391305,0.765826281889366,0.9304347826086956,115.0,981.0,0.2138375095022686,0.5112159174649962,-0.03567863213268257,0.026247204599206077,0.05173538368935207,0.026247204599206077,0.05173538368935207,0.8205128205128205,0.8205128205128205,0.717948717948718,-0.031641816405221804,0.09980551725234024
16
+ ETS2+IKZF3,0.08602150537634409,0.12,0.12,0.08602150537634409,0.08602150537634409,0.17413441955193482,0.12,0.12,0.08,0.072,-0.2000826451113112,0.7526881720430108,0.7392947937511953,0.9193548387096774,186.0,982.0,0.26635620610024097,0.49482180127341413,-0.2771418189558593,0.06653587181552315,0.06513827546247006,0.06653587181552315,0.06513827546247006,0.17948717948717952,0.1282051282051282,0.10256410256410253,-0.031641816405221804,0.09980551725234024
17
+ FEV+ISL2,0.03289473684210526,0.02,0.02,0.03289473684210526,0.03289473684210526,0.13993871297242083,0.02,0.02,0.07,0.072,-0.2000826451113112,0.75,0.7467580642404144,0.9013157894736842,152.0,979.0,0.2536131263493009,0.5152293321747766,-0.13651182007271193,0.04626299293094251,0.06463670563077992,0.04626299293094251,0.06463670563077992,0.23076923076923073,0.3076923076923077,0.5897435897435898,-0.031641816405221804,0.09980551725234024
18
+ FOSB+CEBPB,0.1803921568627451,0.2,0.2,0.2,0.1803921568627451,0.24719101123595505,0.2,0.2,0.2,0.158,-0.2000826451113112,0.8509803921568627,0.7894731508158271,0.9490196078431372,255.0,979.0,0.28944713544660444,0.5101302803000396,0.036243848389008786,0.07338186133634506,0.09196475653884247,0.07338186133634506,0.09196475653884247,0.07692307692307687,0.05128205128205132,0.8461538461538461,-0.031641816405221804,0.09980551725234024
19
+ FOXA3+FOXA1,0.09285714285714286,0.08,0.1,0.09285714285714286,0.09285714285714286,0.13136456211812628,0.08,0.1,0.085,0.07,-0.2000826451113112,0.6785714285714286,0.8088563306363438,0.9214285714285714,140.0,982.0,0.22712627586112877,0.5116279069767441,-0.10712739369482961,0.03451801773627973,0.058775680718583995,0.03451801773627973,0.058775680718583995,0.28205128205128205,0.7435897435897436,0.6153846153846154,-0.031641816405221804,0.09980551725234024
20
+ KLF1+BAK1,0.041666666666666664,0.06,0.041666666666666664,0.041666666666666664,0.041666666666666664,0.08460754332313965,0.06,0.04,0.045,0.044,-0.2000826451113112,0.6666666666666666,0.6905905661401509,0.8645833333333334,96.0,981.0,0.18168786701598455,0.5141961651917405,-0.31031788066882127,0.038295942256511235,0.05514423061684422,0.038295942256511235,0.05514423061684422,0.717948717948718,0.6666666666666667,0.05128205128205132,-0.031641816405221804,0.09980551725234024
21
+ KLF1+CEBPA,0.3203125,0.12,0.31,0.295,0.3203125,0.37741607324516785,0.12,0.31,0.295,0.318,-0.2000826451113112,0.8567708333333334,0.7848429071963489,0.9661458333333334,384.0,983.0,0.43969830739977106,0.536041920319264,0.022255126257146953,0.0750081817556599,0.0953036016207881,0.07500818175565989,0.0953036016207881,0.05128205128205132,0.07692307692307687,0.8461538461538461,-0.031641816405221804,0.09980551725234024
22
+ KLF1+CLDN6,0.04580152671755725,0.04,0.04,0.04580152671755725,0.04580152671755725,0.12016293279022404,0.04,0.04,0.065,0.06,-0.2000826451113112,0.6870229007633588,0.7372944693572496,0.9007633587786259,131.0,982.0,0.28087759295560244,0.5318915310218818,-0.31587954989032513,0.06404557414061014,0.06751865557073668,0.06404557414061014,0.06751865557073668,0.1282051282051282,0.17948717948717952,0.02564102564102566,-0.031641816405221804,0.09980551725234024
23
+ LYL1+CEBPB,0.08860759493670886,0.06,0.07,0.08860759493670886,0.08860759493670886,0.15061475409836064,0.06,0.07,0.1,0.086,-0.2000826451113112,0.7721518987341772,0.868395024908931,0.930379746835443,158.0,976.0,0.2324005330916518,0.520595928921495,-0.13827426473899554,0.0360505190620416,0.06021773479388278,0.0360505190620416,0.06021773479388278,0.33333333333333337,0.7435897435897436,0.4871794871794872,-0.031641816405221804,0.09980551725234024
24
+ MAP2K3+ELMSAN1,0.032467532467532464,0.02,0.02,0.032467532467532464,0.032467532467532464,0.1443089430894309,0.02,0.02,0.04,0.054,-0.2000826451113112,0.7207792207792207,0.732369967871367,0.922077922077922,154.0,984.0,0.27333565791083536,0.517557796813116,-0.04289138553384552,0.028153493837487602,0.05046528814853841,0.028153493837487602,0.05046528814853841,0.8974358974358975,0.8461538461538461,0.7692307692307692,-0.031641816405221804,0.09980551725234024
25
+ MAP2K3+IKZF3,0.07407407407407407,0.12,0.08,0.07407407407407407,0.07407407407407407,0.09959349593495935,0.12,0.08,0.055,0.046,-0.2000826451113112,0.6481481481481481,0.6514341273699562,0.9074074074074074,108.0,984.0,0.2288621815591817,0.5090049410396944,-0.2763191253315642,0.07233352024207221,0.06484376322228856,0.07233352024207221,0.06484376322228856,0.20512820512820518,0.07692307692307687,0.1282051282051282,-0.031641816405221804,0.09980551725234024
26
+ MAP2K3+MAP2K6,0.04225352112676056,0.04,0.04225352112676056,0.04225352112676056,0.04225352112676056,0.0681586978636826,0.04,0.03,0.02,0.026,-0.2000826451113112,0.676056338028169,0.7024968831081309,0.9436619718309859,71.0,983.0,0.16722608252289456,0.5205506450977122,0.13307202739409452,0.022966278226882807,0.04853287749778766,0.022966278226882807,0.04853287749778766,0.9743589743589743,0.9743589743589743,0.9743589743589743,-0.031641816405221804,0.09980551725234024
27
+ MAP2K3+SLC38A2,0.017543859649122806,0.02,0.017543859649122806,0.017543859649122806,0.017543859649122806,0.05188199389623601,0.02,0.04,0.045,0.03,-0.2000826451113112,0.7192982456140351,0.6822727272727273,0.8947368421052632,57.0,983.0,0.17234174566710272,0.5264925303715279,0.30981581535034897,0.02171010429886536,0.04873321986835612,0.02171010429886536,0.04873321986835612,1.0,1.0,1.0,-0.031641816405221804,0.09980551725234024
28
+ MAP2K6+ELMSAN1,0.05813953488372093,0.08,0.05813953488372093,0.05813953488372093,0.05813953488372093,0.08020304568527918,0.08,0.06,0.045,0.042,-0.2000826451113112,0.7441860465116279,0.7541226554904238,0.9186046511627907,86.0,985.0,0.14650110892919355,0.4924685766627653,0.029407502060946642,0.025805122902123088,0.049288783957366376,0.025805122902123088,0.049288783957366376,0.9230769230769231,0.9230769230769231,0.9230769230769231,-0.031641816405221804,0.09980551725234024
29
+ MAPK1+IKZF3,0.07926829268292683,0.08,0.08,0.07926829268292683,0.07926829268292683,0.15376782077393075,0.08,0.08,0.07,0.06,-0.2000826451113112,0.7073170731707317,0.7689767989171468,0.9207317073170732,164.0,982.0,0.28966580299530725,0.5186901913875597,-0.21384405258972097,0.04726938927505059,0.060919073434818234,0.04726938927505059,0.060919073434818234,0.2564102564102564,0.2564102564102564,0.15384615384615385,-0.031641816405221804,0.09980551725234024
30
+ PTPN12+PTPN9,0.04081632653061224,0.06,0.04081632653061224,0.04081632653061224,0.04081632653061224,0.09044715447154472,0.06,0.04,0.03,0.028,-0.2000826451113112,0.7448979591836735,0.6082492634586196,0.9081632653061225,98.0,984.0,0.2483617031883729,0.5505961808226615,-0.19877860618913304,0.04181148408887991,0.052487589400356076,0.04181148408887991,0.052487589400356076,0.6923076923076923,0.4871794871794872,0.2564102564102564,-0.031641816405221804,0.09980551725234024
31
+ PTPN12+UBASH3A,0.034782608695652174,0.04,0.04,0.034782608695652174,0.034782608695652174,0.10703363914373089,0.04,0.04,0.055,0.05,-0.2000826451113112,0.7391304347826086,0.6682385997049355,0.9130434782608695,115.0,981.0,0.21502783452022997,0.4870105625153525,-0.16106452255771544,0.048088372886192915,0.05675571948375862,0.048088372886192915,0.05675571948375862,0.41025641025641024,0.33333333333333337,0.28205128205128205,-0.031641816405221804,0.09980551725234024
32
+ SGK1+S1PR2,0.09230769230769231,0.1,0.09,0.09230769230769231,0.09230769230769231,0.18883248730964466,0.1,0.09,0.095,0.102,-0.2000826451113112,0.7333333333333333,0.7583688681190073,0.9538461538461539,195.0,985.0,0.261465134747528,0.52614110527154,-0.1821107713578707,0.05387832234511301,0.06745853240998974,0.05387832234511301,0.06745853240998974,0.17948717948717952,0.3589743589743589,0.33333333333333337,-0.031641816405221804,0.09980551725234024
33
+ SGK1+TBX2,0.09836065573770492,0.14,0.12,0.09836065573770492,0.09836065573770492,0.11507128309572301,0.14,0.12,0.08,0.048,-0.2000826451113112,0.7131147540983607,0.7474755153915058,0.9262295081967213,122.0,982.0,0.2510835253957017,0.5462956047649276,-0.1356539487566576,0.044212260161360505,0.055892741715334794,0.044212260161360505,0.055892741715334794,0.641025641025641,0.5128205128205128,0.5384615384615384,-0.031641816405221804,0.09980551725234024
34
+ SGK1+TBX3,0.045871559633027525,0.06,0.05,0.045871559633027525,0.045871559633027525,0.10558375634517767,0.06,0.05,0.025,0.032,-0.2000826451113112,0.6788990825688074,0.7066795202165329,0.9541284403669725,109.0,985.0,0.2918634033368116,0.5976688392590536,-0.11598589083125053,0.05394683661706369,0.059579804427312796,0.05394683661706369,0.059579804427312796,0.3846153846153846,0.3589743589743589,0.641025641025641,-0.031641816405221804,0.09980551725234024
35
+ TBX3+TBX2,0.030864197530864196,0.04,0.03,0.030864197530864196,0.030864197530864196,0.15306122448979592,0.04,0.03,0.025,0.046,-0.2000826451113112,0.7469135802469136,0.6769876687696892,0.9259259259259259,162.0,980.0,0.34729712004032,0.5836721765520492,-0.01880227915565069,0.044368545103710806,0.054246974930411994,0.044368545103710806,0.054246974930411994,0.8205128205128205,0.5641025641025641,0.7692307692307692,-0.031641816405221804,0.09980551725234024
36
+ TGFBR2+IGDCC3,0.045871559633027525,0.04,0.05,0.045871559633027525,0.045871559633027525,0.1016260162601626,0.04,0.05,0.025,0.04,-0.2000826451113112,0.6330275229357798,0.6153671478555043,0.9174311926605505,109.0,984.0,0.23279528674164413,0.49660725501704095,-0.16690853413319856,0.036883311422263,0.05622611922851837,0.036883311422263,0.05622611922851837,0.6153846153846154,0.717948717948718,0.5384615384615384,-0.031641816405221804,0.09980551725234024
37
+ TGFBR2+PRTG,0.03418803418803419,0.02,0.03,0.03418803418803419,0.03418803418803419,0.10986775178026449,0.02,0.03,0.06,0.054,-0.2000826451113112,0.6752136752136753,0.7058082889904819,0.9230769230769231,117.0,983.0,0.20744749833987094,0.5051785385873722,-0.1829850690465747,0.04087212126325776,0.055552726399332356,0.04087212126325777,0.055552726399332356,0.5384615384615384,0.5384615384615384,0.46153846153846156,-0.031641816405221804,0.09980551725234024
38
+ UBASH3B+OSR2,0.03816793893129771,0.04,0.04,0.03816793893129771,0.03816793893129771,0.12551020408163266,0.04,0.04,0.04,0.052,-0.2000826451113112,0.6946564885496184,0.8146458711676103,0.9389312977099237,131.0,980.0,0.23560822932383,0.553417545832272,-0.05699398157025002,0.027021718933499927,0.05123470444289772,0.027021718933499927,0.05123470444289772,0.8717948717948718,0.8717948717948718,0.6923076923076923,-0.031641816405221804,0.09980551725234024
39
+ UBASH3B+ZBTB25,0.030927835051546393,0.02,0.030927835051546393,0.030927835051546393,0.030927835051546393,0.08952187182095625,0.02,0.03,0.04,0.03,-0.2000826451113112,0.7216494845360825,0.6057634764995626,0.9072164948453608,97.0,983.0,0.19439001023632266,0.4971629505314473,-0.15276859081569982,0.04218872739041526,0.05453678239339361,0.04218872739041526,0.05453678239339361,0.6923076923076923,0.4358974358974359,0.3846153846153846,-0.031641816405221804,0.09980551725234024
40
+ ZC3HAV1+CEBPE,0.08024691358024691,0.06,0.1,0.08024691358024691,0.08024691358024691,0.16024340770791076,0.06,0.1,0.065,0.074,-0.2000826451113112,0.7777777777777778,0.8506234847083509,0.9753086419753086,162.0,986.0,0.23160958067296245,0.5185627154600901,0.09038161943839848,0.024524203594263903,0.05225556549413856,0.024524203594263903,0.05225556549413856,0.8717948717948718,0.8974358974358975,0.9487179487179487,-0.031641816405221804,0.09980551725234024
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/agg_results.csv ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ statistic,overlap_at_N,overlap_at_50,overlap_at_100,overlap_at_200,overlap_at_500,precision_at_N,precision_at_50,precision_at_100,precision_at_200,precision_at_500,de_spearman_sig,de_direction_match,de_spearman_lfc_sig,de_sig_genes_recall,de_nsig_counts_real,de_nsig_counts_pred,pr_auc,roc_auc,pearson_delta,mse,mae,mse_delta,mae_delta,discrimination_score_l1,discrimination_score_l2,discrimination_score_cosine,pearson_edistance,clustering_agreement
2
+ count,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0,39.0
3
+ null_count,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
4
+ mean,0.07395320375987084,0.06849358974358978,0.0725404899006272,0.07431250785187005,0.07395320375987084,0.12877704205127136,0.06871794871794874,0.07282051282051284,0.07500000000000001,0.06682051282051284,0.0006180401974148523,0.7319719710281621,0.7355229056966586,0.9061342423642376,137.89743589743588,977.6153846153846,0.21954205885448827,0.502651350313865,-0.08164090049789667,0.03866601647089939,0.054930761086430545,0.03866601647089939,0.054930761086430545,0.5141354372123603,0.5121630506245891,0.5115055884286653,0.12324841081988779,0.007165149729861788
5
+ std,0.06327073571923675,0.04142623783921817,0.06430260938703716,0.06497823680683154,0.06327073571923675,0.0713575372552735,0.04124251046390996,0.06398802182105803,0.06369127842466676,0.06055657285150447,1.0983753514559728e-19,0.05548309801759892,0.06748831752084138,0.027067878739138167,72.75871981447884,2.2197129332580094,0.07449489082434584,0.025106101263114602,0.11079253285523155,0.01804318670820533,0.013893820583993645,0.01804318670820533,0.013893820583993646,0.2932360142364609,0.29232232244842243,0.30752068526485193,2.8118408997272905e-17,2.636100843494335e-18
6
+ min,0.0,0.0,0.0,0.0,0.0,0.030706243602865915,0.0,0.0,0.005,0.008,0.0006180401974148526,0.6301369863013698,0.5466766243465273,0.8247422680412371,32.0,972.0,0.07528683946739989,0.45502409257502185,-0.25674855274153263,0.017089386967776357,0.041686743621341026,0.017089386967776357,0.041686743621341026,0.02564102564102566,0.02564102564102566,0.02564102564102566,0.12324841081988787,0.007165149729861791
7
+ 25%,0.042105263157894736,0.04,0.04,0.042105263157894736,0.042105263157894736,0.09397344228804903,0.04,0.04,0.045,0.038,0.0006180401974148526,0.7037037037037037,0.6922532601004144,0.8904109589041096,103.0,976.0,0.17885902518604974,0.48868455014749274,-0.16340383087297497,0.029862651934889305,0.048704704053449004,0.029862651934889305,0.048704704053449004,0.3076923076923077,0.3076923076923077,0.2564102564102564,0.12324841081988787,0.007165149729861791
8
+ 50%,0.05454545454545454,0.06,0.0547945205479452,0.05454545454545454,0.05454545454545454,0.10827374872318693,0.06,0.05,0.055,0.048,0.0006180401974148526,0.7272727272727273,0.7283894176721877,0.9029126213592233,117.0,978.0,0.2155530179012921,0.5015608589100246,-0.10351595518920989,0.03463771528656705,0.05123648361160402,0.03463771528656705,0.05123648361160402,0.5128205128205128,0.5384615384615384,0.5128205128205128,0.12324841081988787,0.007165149729861791
9
+ 75%,0.08333333333333333,0.1,0.08,0.08333333333333333,0.08333333333333333,0.14974358974358976,0.1,0.08,0.08,0.068,0.0006180401974148526,0.7551020408163265,0.793306240068219,0.9259259259259259,158.0,979.0,0.2577585712884973,0.5154239019407559,0.007450254288251119,0.04417537517338548,0.0574020448039485,0.04417537517338548,0.0574020448039485,0.7435897435897436,0.7692307692307692,0.7948717948717949,0.12324841081988787,0.007165149729861791
10
+ max,0.30548302872062666,0.16,0.34,0.33,0.30548302872062666,0.3719262295081967,0.16,0.34,0.33,0.3,0.0006180401974148526,0.8705882352941177,0.8543173721436267,0.9506172839506173,384.0,983.0,0.41857905891786573,0.5534962262791009,0.22806546181808088,0.11179507217860543,0.1123107911192389,0.11179507217860543,0.1123107911192389,1.0,1.0,1.0,0.12324841081988787,0.007165149729861791
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ea82f6a522f226c5b36c53fc2e7f2cab3036a82f79aa65f82d0553b53de5342
3
+ size 61463408
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/pred.h5ad ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d77d0b092677a1ffcf600ebedc1aa980f894e98d68dd9a9a75e4a00b1d5a8af2
3
+ size 49952825
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/real.h5ad ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d906eb6262ae9d587a091c8f6722e562182c9cd62e18f7c2faecbd7c4935ec3c
3
+ size 75069953
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/iteration_190000/results.csv ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ perturbation,overlap_at_N,overlap_at_50,overlap_at_100,overlap_at_200,overlap_at_500,precision_at_N,precision_at_50,precision_at_100,precision_at_200,precision_at_500,de_spearman_sig,de_direction_match,de_spearman_lfc_sig,de_sig_genes_recall,de_nsig_counts_real,de_nsig_counts_pred,pr_auc,roc_auc,pearson_delta,mse,mae,mse_delta,mae_delta,discrimination_score_l1,discrimination_score_l2,discrimination_score_cosine,pearson_edistance,clustering_agreement
2
+ AHR+FEV,0.1308411214953271,0.14,0.14,0.125,0.1308411214953271,0.20245398773006135,0.14,0.14,0.125,0.1,0.0006180401974148526,0.7897196261682243,0.793306240068219,0.9252336448598131,214.0,978.0,0.289355620021763,0.5044885971796151,-0.000681255195205219,0.05580747882233428,0.0703616503153296,0.05580747882233427,0.0703616503153296,0.10256410256410253,0.15384615384615385,0.7948717948717949,0.12324841081988787,0.007165149729861791
3
+ AHR+KLF1,0.1,0.1,0.09,0.1,0.1,0.12589559877175024,0.1,0.09,0.085,0.068,0.0006180401974148526,0.6714285714285714,0.7230217243868823,0.8785714285714286,140.0,977.0,0.2155530179012921,0.5011627906976744,-0.17882777360167185,0.03415252740506693,0.053647045372835565,0.03415252740506693,0.053647045372835565,0.3076923076923077,0.46153846153846156,0.10256410256410253,0.12324841081988787,0.007165149729861791
4
+ BCL2L11+BAK1,0.03125,0.03125,0.03125,0.03125,0.03125,0.030706243602865915,0.04,0.05,0.045,0.03,0.0006180401974148526,0.75,0.8320806599450046,0.9375,32.0,977.0,0.07528683946739989,0.5477143595041322,0.007450254288251119,0.018631415934436527,0.042142537012775594,0.018631415934436527,0.042142537012775594,0.9487179487179487,0.9487179487179487,0.7948717948717949,0.12324841081988787,0.007165149729861791
5
+ BPGM+ZBTB1,0.05454545454545454,0.08,0.06,0.05454545454545454,0.05454545454545454,0.10601427115188583,0.08,0.06,0.05,0.04,0.0006180401974148526,0.7272727272727273,0.7216076912323739,0.9454545454545454,110.0,981.0,0.15723145661369453,0.4712257405515833,-0.14295966613751018,0.03777198411559816,0.05109172018229229,0.03777198411559816,0.05109172018229229,0.5128205128205128,0.3846153846153846,0.33333333333333337,0.12324841081988787,0.007165149729861791
6
+ CBL+PTPN12,0.04807692307692308,0.06,0.05,0.04807692307692308,0.04807692307692308,0.09397344228804903,0.06,0.05,0.05,0.03,0.0006180401974148526,0.7115384615384616,0.5466766243465273,0.8846153846153846,104.0,979.0,0.2527906962637153,0.5022428743131868,-0.18550991579880255,0.04731173214331579,0.05330420638914829,0.04731173214331579,0.05330420638914829,0.3846153846153846,0.23076923076923073,0.17948717948717952,0.12324841081988787,0.007165149729861791
7
+ CBL+PTPN9,0.042105263157894736,0.02,0.042105263157894736,0.042105263157894736,0.042105263157894736,0.08709016393442623,0.02,0.04,0.05,0.032,0.0006180401974148526,0.7157894736842105,0.6192082645090848,0.8947368421052632,95.0,976.0,0.23347166134256753,0.5006746147135795,-0.11418825671976725,0.0345185799724352,0.048704704053449004,0.0345185799724352,0.048704704053449004,0.717948717948718,0.5897435897435898,0.46153846153846156,0.12324841081988787,0.007165149729861791
8
+ CBL+TGFBR2,0.0547945205479452,0.04,0.0547945205479452,0.0547945205479452,0.0547945205479452,0.06659836065573771,0.04,0.04,0.05,0.038,0.0006180401974148526,0.6301369863013698,0.6730638691761802,0.8904109589041096,73.0,976.0,0.1657458038482859,0.5353327126834242,-0.10350698889045816,0.03266801215509963,0.04836069691089276,0.03266801215509963,0.04836069691089276,0.7435897435897436,0.641025641025641,0.5897435897435898,0.12324841081988787,0.007165149729861791
9
+ CBL+UBASH3A,0.0,0.0,0.0,0.0,0.0,0.043209876543209874,0.0,0.0,0.005,0.008,0.0006180401974148526,0.6595744680851063,0.8201896392229417,0.8936170212765957,47.0,972.0,0.12664611232108428,0.5477662923355139,-0.16340383087297497,0.03660779070416713,0.05174531782519213,0.03660779070416713,0.05174531782519213,0.5128205128205128,0.5384615384615384,0.2564102564102564,0.12324841081988787,0.007165149729861791
10
+ CBL+UBASH3B,0.08904109589041095,0.14,0.1,0.08904109589041095,0.08904109589041095,0.13190184049079753,0.14,0.1,0.07,0.058,0.0006180401974148526,0.7328767123287672,0.6719708226394705,0.8835616438356164,146.0,978.0,0.2927739461231338,0.5109236148984634,-0.09303837127636658,0.047811851528487485,0.05215327697445609,0.047811851528487485,0.05215327697445609,0.3589743589743589,0.20512820512820518,0.4871794871794872,0.12324841081988787,0.007165149729861791
11
+ CDKN1B+CDKN1A,0.02912621359223301,0.02,0.03,0.02912621359223301,0.02912621359223301,0.09489795918367347,0.02,0.03,0.06,0.048,0.0006180401974148526,0.7572815533980582,0.7994003426613364,0.9029126213592233,103.0,980.0,0.13727131313018917,0.4975701096427142,-0.15707749719183386,0.029985238722000304,0.05206961946948075,0.029985238722000304,0.05206961946948075,0.46153846153846156,0.641025641025641,0.33333333333333337,0.12324841081988787,0.007165149729861791
12
+ CDKN1C+CDKN1B,0.03636363636363636,0.04,0.03,0.03636363636363636,0.03636363636363636,0.1023541453428864,0.04,0.03,0.045,0.042,0.0006180401974148526,0.7090909090909091,0.8149440522239354,0.9090909090909091,110.0,977.0,0.15180512654084866,0.5154239019407559,-0.14115175381435643,0.029967122965036647,0.05138300370287098,0.029967122965036647,0.05138300370287098,0.5897435897435898,0.6923076923076923,0.5128205128205128,0.12324841081988787,0.007165149729861791
13
+ CEBPE+CEBPA,0.30548302872062666,0.1,0.34,0.33,0.30548302872062666,0.3707865168539326,0.1,0.34,0.33,0.3,0.0006180401974148526,0.8694516971279374,0.729553310592033,0.9477806788511749,383.0,979.0,0.4157540694520757,0.49803648581741855,0.04068628601729422,0.11179507217860543,0.1123107911192389,0.11179507217860543,0.1123107911192389,0.02564102564102566,0.02564102564102566,0.9230769230769231,0.12324841081988787,0.007165149729861791
14
+ CEBPE+CNN1,0.11564625850340136,0.14,0.12,0.11564625850340136,0.11564625850340136,0.14096016343207354,0.14,0.12,0.125,0.082,0.0006180401974148526,0.7551020408163265,0.8324352511733009,0.9387755102040817,147.0,979.0,0.23976185026595853,0.5499198507069885,-0.04180838169993791,0.02799969093443837,0.05105374262657179,0.02799969093443837,0.05105374262657179,0.5897435897435898,0.7948717948717949,0.6923076923076923,0.12324841081988787,0.007165149729861791
15
+ ETS2+IGDCC3,0.06956521739130435,0.06,0.06,0.06956521739130435,0.06956521739130435,0.10633946830265849,0.06,0.06,0.075,0.058,0.0006180401974148526,0.7391304347826086,0.7768612654710786,0.9043478260869565,115.0,978.0,0.23370533925403658,0.495897813804962,0.029532205030916036,0.020717920454122067,0.045982380799085745,0.020717920454122067,0.045982380799085745,0.8205128205128205,0.8205128205128205,0.8205128205128205,0.12324841081988787,0.007165149729861791
16
+ ETS2+IKZF3,0.10752688172043011,0.12,0.11,0.10752688172043011,0.10752688172043011,0.17008196721311475,0.12,0.11,0.1,0.086,0.0006180401974148526,0.7634408602150538,0.7704231330786419,0.8924731182795699,186.0,976.0,0.2453396755066589,0.48778764101344746,-0.18685575986849334,0.05693390716612494,0.060505517368878346,0.05693390716612494,0.060505517368878346,0.1282051282051282,0.1282051282051282,0.07692307692307687,0.12324841081988787,0.007165149729861791
17
+ FEV+ISL2,0.05263157894736842,0.04,0.04,0.05263157894736842,0.05263157894736842,0.13655030800821355,0.04,0.04,0.065,0.066,0.0006180401974148526,0.7105263157894737,0.7411408742029778,0.875,152.0,974.0,0.1802385047469924,0.46315246400198606,-0.08184300427470183,0.04140964303946295,0.05993504252580628,0.04140964303946295,0.05993504252580628,0.20512820512820518,0.2564102564102564,0.41025641025641024,0.12324841081988787,0.007165149729861791
18
+ FOSB+CEBPB,0.1803921568627451,0.16,0.19,0.185,0.1803921568627451,0.24487704918032788,0.16,0.19,0.185,0.15,0.0006180401974148526,0.8705882352941177,0.7788869424215662,0.9372549019607843,255.0,976.0,0.2895674815827421,0.49720226345571794,0.07216521168170856,0.06788741332834797,0.08781863929818466,0.06788741332834797,0.08781863929818468,0.07692307692307687,0.05128205128205132,0.8461538461538461,0.12324841081988787,0.007165149729861791
19
+ FOXA3+FOXA1,0.07857142857142857,0.1,0.08,0.07857142857142857,0.07857142857142857,0.12883435582822086,0.1,0.08,0.075,0.064,0.0006180401974148526,0.6857142857142857,0.8037918215613383,0.9,140.0,978.0,0.17885902518604974,0.4941901993355481,-0.05074626553588181,0.029862651934889305,0.05398229729784066,0.029862651934889305,0.05398229729784066,0.33333333333333337,0.7692307692307692,0.641025641025641,0.12324841081988787,0.007165149729861791
20
+ KLF1+BAK1,0.052083333333333336,0.1,0.052083333333333336,0.052083333333333336,0.052083333333333336,0.08597748208802457,0.1,0.05,0.05,0.042,0.0006180401974148526,0.7083333333333334,0.6907398360744721,0.875,96.0,977.0,0.19735133996261323,0.48868455014749274,-0.24515315071378704,0.03197165400949463,0.04877062774155349,0.03197165400949463,0.04877062774155349,0.7435897435897436,0.6666666666666667,0.02564102564102566,0.12324841081988787,0.007165149729861791
21
+ KLF1+CEBPA,0.2942708333333333,0.1,0.25,0.285,0.2942708333333333,0.3719262295081967,0.1,0.25,0.285,0.298,0.0006180401974148526,0.8671875,0.7524491829127294,0.9453125,384.0,976.0,0.41857905891786573,0.5074277935606061,0.033241542515528716,0.06595129422363945,0.08905813105074563,0.06595129422363945,0.08905813105074563,0.05128205128205132,0.07692307692307687,0.8717948717948718,0.12324841081988787,0.007165149729861791
22
+ KLF1+CLDN6,0.05343511450381679,0.04,0.04,0.05343511450381679,0.05343511450381679,0.12040816326530612,0.04,0.04,0.055,0.056,0.0006180401974148526,0.7022900763358778,0.7190155882980995,0.9007633587786259,131.0,980.0,0.21494914880766414,0.5019984363882325,-0.25674855274153263,0.051408548351699966,0.059620483832245184,0.051408548351699966,0.059620483832245184,0.17948717948717952,0.17948717948717952,0.05128205128205132,0.12324841081988787,0.007165149729861791
23
+ LYL1+CEBPB,0.0949367088607595,0.06,0.08,0.0949367088607595,0.0949367088607595,0.14974358974358976,0.06,0.08,0.095,0.078,0.0006180401974148526,0.7848101265822784,0.8251475211277035,0.9240506329113924,158.0,975.0,0.218219692606571,0.504111669021919,-0.16514919028268413,0.029257650899615014,0.053947390749385,0.029257650899615014,0.053947390749385,0.3076923076923077,0.6666666666666667,0.1282051282051282,0.12324841081988787,0.007165149729861791
24
+ MAP2K3+ELMSAN1,0.025974025974025976,0.0,0.02,0.025974025974025976,0.025974025974025976,0.14504596527068436,0.0,0.02,0.04,0.06,0.0006180401974148526,0.7272727272727273,0.718742031589043,0.922077922077922,154.0,979.0,0.2577585712884973,0.48963418378311996,0.020462538397928824,0.019811996728230724,0.04367858584334204,0.019811996728230724,0.04367858584334204,0.8974358974358975,0.8461538461538461,0.8205128205128205,0.12324841081988787,0.007165149729861791
25
+ MAP2K3+IKZF3,0.08333333333333333,0.1,0.08,0.08333333333333333,0.08333333333333333,0.09805924412665985,0.1,0.08,0.055,0.046,0.0006180401974148526,0.7037037037037037,0.7167014593878388,0.8888888888888888,108.0,979.0,0.22899973834921372,0.4874553645573824,-0.20767238076053746,0.0652399118899808,0.06086253874483393,0.0652399118899808,0.06086253874483393,0.20512820512820518,0.10256410256410253,0.1282051282051282,0.12324841081988787,0.007165149729861791
26
+ MAP2K3+MAP2K6,0.028169014084507043,0.04,0.028169014084507043,0.028169014084507043,0.028169014084507043,0.065439672801636,0.04,0.03,0.025,0.034,0.0006180401974148526,0.6901408450704225,0.6922532601004144,0.9014084507042254,71.0,978.0,0.130106127841948,0.5246061947573492,0.10542488939536575,0.017089386967776357,0.042199752748710434,0.017089386967776357,0.042199752748710434,0.9743589743589743,0.9743589743589743,0.9743589743589743,0.12324841081988787,0.007165149729861791
27
+ MAP2K3+SLC38A2,0.05263157894736842,0.04,0.05263157894736842,0.05263157894736842,0.05263157894736842,0.051934826883910386,0.04,0.05,0.035,0.032,0.0006180401974148526,0.7017543859649122,0.6621428571428571,0.8947368421052632,57.0,982.0,0.09447034806783824,0.45502409257502185,0.22806546181808088,0.01764320931428523,0.041686743621341026,0.01764320931428523,0.041686743621341026,1.0,1.0,1.0,0.12324841081988787,0.007165149729861791
28
+ MAP2K6+ELMSAN1,0.046511627906976744,0.02,0.046511627906976744,0.046511627906976744,0.046511627906976744,0.07865168539325842,0.02,0.05,0.055,0.046,0.0006180401974148526,0.7558139534883721,0.7393594426981565,0.8953488372093024,86.0,979.0,0.14357163096260944,0.46603226298916084,0.07308970049844386,0.01884736438588656,0.04362012745480605,0.01884736438588656,0.04362012745480605,0.9487179487179487,0.9230769230769231,0.9487179487179487,0.12324841081988787,0.007165149729861791
29
+ MAPK1+IKZF3,0.07317073170731707,0.04,0.04,0.07317073170731707,0.07317073170731707,0.15462868769074262,0.04,0.04,0.075,0.058,0.0006180401974148526,0.725609756097561,0.7737163223801006,0.926829268292683,164.0,983.0,0.27474306161131856,0.5015608589100246,-0.18217857153146724,0.0425078665820119,0.0574020448039485,0.0425078665820119,0.0574020448039485,0.2564102564102564,0.3076923076923077,0.1282051282051282,0.12324841081988787,0.007165149729861791
30
+ PTPN12+PTPN9,0.061224489795918366,0.06,0.061224489795918366,0.061224489795918366,0.061224489795918366,0.0869120654396728,0.06,0.06,0.045,0.03,0.0006180401974148526,0.7142857142857143,0.6621602663027536,0.8673469387755102,98.0,978.0,0.25011674665005473,0.5097402597402598,-0.20268265319102213,0.03914455888068411,0.05023655449671707,0.03914455888068411,0.05023655449671707,0.6923076923076923,0.4358974358974359,0.20512820512820518,0.12324841081988787,0.007165149729861791
31
+ PTPN12+UBASH3A,0.06086956521739131,0.1,0.07,0.06086956521739131,0.06086956521739131,0.10245901639344263,0.1,0.07,0.065,0.048,0.0006180401974148526,0.7304347826086957,0.6872912173681067,0.8695652173913043,115.0,976.0,0.20163301499093328,0.4667747482191108,-0.1405571263181629,0.04187065351581521,0.05191913093302079,0.041870653515815204,0.05191913093302079,0.41025641025641024,0.33333333333333337,0.2564102564102564,0.12324841081988787,0.007165149729861791
32
+ SGK1+S1PR2,0.08205128205128205,0.08,0.07,0.08205128205128205,0.08205128205128205,0.18692543411644535,0.08,0.07,0.085,0.104,0.0006180401974148526,0.7487179487179487,0.7726498750939433,0.9384615384615385,195.0,979.0,0.27246186996366134,0.504405160057334,-0.1425496176526846,0.04362755416454579,0.061769647398377175,0.04362755416454579,0.061769647398377175,0.15384615384615385,0.3589743589743589,0.3589743589743589,0.12324841081988787,0.007165149729861791
33
+ SGK1+TBX2,0.06557377049180328,0.1,0.06,0.06557377049180328,0.06557377049180328,0.11349693251533742,0.1,0.06,0.065,0.052,0.0006180401974148526,0.7295081967213115,0.7230355295385358,0.9098360655737705,122.0,978.0,0.23721339749548936,0.5243194294036371,-0.17427454250530847,0.04053582957303282,0.0501051497886934,0.04053582957303282,0.0501051497886934,0.5641025641025641,0.41025641025641024,0.20512820512820518,0.12324841081988787,0.007165149729861791
34
+ SGK1+TBX3,0.045871559633027525,0.08,0.05,0.045871559633027525,0.045871559633027525,0.10245901639344263,0.08,0.05,0.025,0.024,0.0006180401974148526,0.6880733944954128,0.6942678111269721,0.9174311926605505,109.0,976.0,0.3383204059729327,0.5534962262791009,-0.08464092901325047,0.04417537517338548,0.05123648361160402,0.04417537517338548,0.05123648361160402,0.5641025641025641,0.3076923076923077,0.6666666666666667,0.12324841081988787,0.007165149729861791
35
+ TBX3+TBX2,0.030864197530864196,0.02,0.03,0.030864197530864196,0.030864197530864196,0.15337423312883436,0.02,0.03,0.03,0.066,0.0006180401974148526,0.7530864197530864,0.7208131055895328,0.9259259259259259,162.0,978.0,0.2846346150206432,0.5430920180323522,-0.019667607914308216,0.03463771528656705,0.04748869970105704,0.03463771528656705,0.04748869970105704,0.7948717948717949,0.5384615384615384,0.7692307692307692,0.12324841081988787,0.007165149729861791
36
+ TGFBR2+IGDCC3,0.03669724770642202,0.06,0.04,0.03669724770642202,0.03669724770642202,0.09908069458631256,0.06,0.04,0.04,0.03,0.0006180401974148526,0.6330275229357798,0.6363139235813425,0.8899082568807339,109.0,979.0,0.18103811865440816,0.4730485280943997,-0.10351595518920989,0.031240376498607152,0.05090387243412622,0.031240376498607152,0.05090387243412622,0.6153846153846154,0.7692307692307692,0.6153846153846154,0.12324841081988787,0.007165149729861791
37
+ TGFBR2+PRTG,0.03418803418803419,0.04,0.04,0.03418803418803419,0.03418803418803419,0.10827374872318693,0.04,0.04,0.045,0.046,0.0006180401974148526,0.6752136752136753,0.7283894176721877,0.905982905982906,117.0,979.0,0.1710995969889653,0.4782598174444153,-0.14178338892395337,0.03324635981127433,0.05022525364973255,0.03324635981127433,0.05022525364973255,0.46153846153846156,0.5641025641025641,0.5128205128205128,0.12324841081988787,0.007165149729861791
38
+ UBASH3B+OSR2,0.04580152671755725,0.1,0.06,0.04580152671755725,0.04580152671755725,0.12397540983606557,0.1,0.06,0.05,0.046,0.0006180401974148526,0.732824427480916,0.8088719153936546,0.9236641221374046,131.0,976.0,0.18232471960433552,0.5158161965582972,-0.02485046822501524,0.02185241151506355,0.044643295457333705,0.02185241151506355,0.044643295457333705,0.8974358974358975,0.8717948717948718,0.717948717948718,0.12324841081988787,0.007165149729861791
39
+ UBASH3B+ZBTB25,0.010309278350515464,0.02,0.010309278350515464,0.010309278350515464,0.010309278350515464,0.08221993833504625,0.02,0.02,0.035,0.032,0.0006180401974148526,0.711340206185567,0.6584528577347248,0.8247422680412371,97.0,973.0,0.18033908245418523,0.47326209313742285,-0.1533095016375446,0.036293167827057715,0.04955475749838851,0.036293167827057715,0.04955475749838851,0.6666666666666667,0.41025641025641024,0.3589743589743589,0.12324841081988787,0.007165149729861791
40
+ ZC3HAV1+CEBPE,0.08024691358024691,0.08,0.08,0.08024691358024691,0.08024691358024691,0.15778688524590165,0.08,0.08,0.08,0.078,0.0006180401974148526,0.8148148148148148,0.8543173721436267,0.9506172839506173,162.0,976.0,0.20305246954480483,0.5139404519873891,0.09221914841694258,0.019781723262453504,0.0468182215665005,0.019781723262453504,0.0468182215665005,0.8461538461538461,0.8974358974358975,0.9487179487179487,0.12324841081988787,0.007165149729861791
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/loss_curve.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20591f9d19719429e00b677f9cd3986ab44ab02e9b866914e1700c8cf579c7d5
3
+ size 16264567
transfer/code/CCFM/result/ccfm-norman-f1-topk30-negTrue-d128-lr5e-05-lw1.0-lp0.4-ema0.9999-ln-wu2000-rk4-online/tb_logs/events.out.tfevents.1773517868.b0019.200.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:371aa81e99d9b9d674bb8602fa2352d0784b6397d464772ee5416188b22342e5
3
+ size 88