export TORCH_NCCL_TIMEOUT=1800000 accelerate launch --config_file configs/accelerate_config.yaml scripts/train.py \ --dataset_base_path "/mnt/nfs/workspace/syq/dataset/Hypersim/processed_normal,/mnt/nfs/workspace/syq/dataset/InteriorVerse/processed_normal,/mnt/nfs/workspace/syq/dataset/sintel" \ --dataset_metadata_path "./data_split/hypersim_normals/hypersim_filtered_all_checked.txt,./data_split/interiorverse_normals/interiorverse_filtered_all.txt,./data_split/sintel_normals/sintel_filtered.txt" \ --data_file_keys "kontext_images,image" \ --model_paths "./FLUX.1-Kontext-dev" \ --learning_rate "1e-5" \ --num_epochs "8" \ --remove_prefix_in_ckpt "pipe.dit." \ --trainable_models "dit" \ --extra_inputs "kontext_images" \ --use_gradient_checkpointing \ --default_caption "Transform to normal map while maintaining original composition" \ --batch_size "8" \ --output_path "ckpts/kontext_normal/bs16_lora" \ --eval_file_list "./data_split/nyu_normals/nyuv2_test2.txt" \ --multi_res_noise \ --save_steps "200" \ --eval_steps "50" \ --with_mask \ --lora_base_model "dit" \ --lora_target_modules "a_to_qkv,b_to_qkv,ff_a.0,ff_a.2,ff_b.0,ff_b.2,a_to_out,b_to_out,proj_out,norm.linear,norm1_a.linear,norm1_b.linear,to_qkv_mlp" \ --lora_rank 64 \ --align_to_opensource_format \ --resume \ --height 768 \ --width 768 \ --adamw8bit # --extra_loss "cycle_consistency_normal_estimation" \ # --deterministic_flow