nohup: ignoring input 检测到 4 个GPU 每个GPU批次大小: 4 总有效批次大小: 16 ===== SD3 LoRA 多GPU训练开始 ===== 模型: /gemini/space/hsd/project/pretrained_model/huggingface/hub/models--stabilityai--stable-diffusion-3-medium-diffusers/snapshots/ea42f8cef0f178587cf766dc8129abd379c90671 输出目录: sd3-lora-finetuned-batch-8 分辨率: 512 每个GPU批次大小: 4 梯度累积步数: 1 总有效批次大小: 16 学习率: 1e-5 最大训练步数: 500000 LoRA Rank: 32 使用GPU: 0,1,2,3 断点重训: latest =========================================== 使用 accelerate 启动多GPU训练... /root/miniconda3/envs/SiT/lib/python3.10/site-packages/transformers/utils/hub.py:111: FutureWarning: Using `TRANSFORMERS_CACHE` is deprecated and will be removed in v5 of Transformers. Use `HF_HOME` instead. warnings.warn( Terminated =========================================== 训练完成! 模型保存在: sd3-lora-finetuned-batch-8 日志保存在: sd3-lora-finetuned-batch-8/logs 验证图片保存在: sd3-lora-finetuned-batch-8/validation_images ===========================================