| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """ |
| Fine-tune Vision Language Models using streaming datasets and Unsloth optimizations. |
| |
| Streams data directly from the Hub - no disk space needed for massive VLM datasets. |
| Uses Unsloth for ~60% less VRAM and 2x faster training. |
| |
| Run locally (if you have a GPU): |
| uv run vlm-streaming-sft-unsloth.py \ |
| --max-steps 100 \ |
| --output-repo your-username/vlm-test |
| |
| Run on HF Jobs: |
| hf jobs uv run vlm-streaming-sft-unsloth.py \ |
| --flavor a100-large \ |
| --secrets HF_TOKEN \ |
| -- \ |
| --max-steps 500 \ |
| --output-repo your-username/vlm-finetuned |
| |
| With Trackio dashboard: |
| uv run vlm-streaming-sft-unsloth.py \ |
| --max-steps 500 \ |
| --output-repo your-username/vlm-finetuned \ |
| --trackio-space your-username/trackio |
| """ |
|
|
| import argparse |
| import logging |
| import os |
| import sys |
| import time |
|
|
| logging.basicConfig( |
| level=logging.INFO, |
| format="%(asctime)s - %(levelname)s - %(message)s", |
| ) |
| logger = logging.getLogger(__name__) |
|
|
|
|
| def check_cuda(): |
| """Check CUDA availability and exit if not available.""" |
| import torch |
|
|
| if not torch.cuda.is_available(): |
| logger.error("CUDA is not available. This script requires a GPU.") |
| logger.error("Run on a machine with a CUDA-capable GPU or use HF Jobs:") |
| logger.error( |
| " hf jobs uv run vlm-streaming-sft-unsloth.py --flavor a100-large ..." |
| ) |
| sys.exit(1) |
| logger.info(f"CUDA available: {torch.cuda.get_device_name(0)}") |
|
|
|
|
| def parse_args(): |
| parser = argparse.ArgumentParser( |
| description="Fine-tune VLMs with streaming datasets using Unsloth", |
| formatter_class=argparse.RawDescriptionHelpFormatter, |
| epilog=""" |
| Examples: |
| # Quick test run |
| uv run vlm-streaming-sft-unsloth.py \\ |
| --max-steps 50 \\ |
| --output-repo username/vlm-test |
| |
| # Full training with Trackio monitoring |
| uv run vlm-streaming-sft-unsloth.py \\ |
| --max-steps 500 \\ |
| --output-repo username/vlm-finetuned \\ |
| --trackio-space username/trackio |
| |
| # Custom dataset and model |
| uv run vlm-streaming-sft-unsloth.py \\ |
| --base-model unsloth/Qwen3-VL-8B-Instruct-unsloth-bnb-4bit \\ |
| --dataset your-username/your-vlm-dataset \\ |
| --max-steps 1000 \\ |
| --output-repo username/custom-vlm |
| """, |
| ) |
|
|
| |
| parser.add_argument( |
| "--base-model", |
| default="unsloth/gemma-3-4b-pt", |
| help="Base VLM model (default: unsloth/gemma-3-4b-pt)", |
| ) |
| parser.add_argument( |
| "--dataset", |
| default="davanstrien/iconclass-vlm-sft", |
| help="Dataset with 'images' and 'messages' columns (default: davanstrien/iconclass-vlm-sft)", |
| ) |
| parser.add_argument( |
| "--output-repo", |
| required=True, |
| help="HF Hub repo to push model to (e.g., 'username/vlm-finetuned')", |
| ) |
|
|
| |
| parser.add_argument( |
| "--max-steps", |
| type=int, |
| default=500, |
| help="Training steps (default: 500). Required for streaming datasets.", |
| ) |
| parser.add_argument( |
| "--batch-size", |
| type=int, |
| default=2, |
| help="Per-device batch size (default: 2)", |
| ) |
| parser.add_argument( |
| "--gradient-accumulation", |
| type=int, |
| default=4, |
| help="Gradient accumulation steps (default: 4). Effective batch = batch-size * this", |
| ) |
| parser.add_argument( |
| "--learning-rate", |
| type=float, |
| default=2e-4, |
| help="Learning rate (default: 2e-4)", |
| ) |
| parser.add_argument( |
| "--max-seq-length", |
| type=int, |
| default=2048, |
| help="Maximum sequence length (default: 2048)", |
| ) |
|
|
| |
| parser.add_argument( |
| "--lora-r", |
| type=int, |
| default=16, |
| help="LoRA rank (default: 16). Higher = more capacity but more VRAM", |
| ) |
| parser.add_argument( |
| "--lora-alpha", |
| type=int, |
| default=32, |
| help="LoRA alpha (default: 32). Usually 2*r", |
| ) |
|
|
| |
| parser.add_argument( |
| "--trackio-space", |
| default=None, |
| help="HF Space for Trackio dashboard (e.g., 'username/trackio')", |
| ) |
| parser.add_argument( |
| "--save-local", |
| default="vlm-streaming-output", |
| help="Local directory to save model (default: vlm-streaming-output)", |
| ) |
|
|
| return parser.parse_args() |
|
|
|
|
| def main(): |
| args = parse_args() |
|
|
| print("=" * 70) |
| print("VLM Streaming Fine-tuning with Unsloth") |
| print("=" * 70) |
| print("\nConfiguration:") |
| print(f" Base model: {args.base_model}") |
| print(f" Dataset: {args.dataset}") |
| print(f" Max steps: {args.max_steps}") |
| print( |
| f" Batch size: {args.batch_size} x {args.gradient_accumulation} = {args.batch_size * args.gradient_accumulation}" |
| ) |
| print(f" Learning rate: {args.learning_rate}") |
| print(f" LoRA rank: {args.lora_r}") |
| print(f" Output repo: {args.output_repo}") |
| print(f" Trackio space: {args.trackio_space or '(not configured)'}") |
| print() |
|
|
| |
| check_cuda() |
|
|
| |
| os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1" |
|
|
| |
| if args.trackio_space: |
| os.environ["TRACKIO_SPACE_ID"] = args.trackio_space |
| logger.info(f"Trackio dashboard: https://huggingface.co/spaces/{args.trackio_space}") |
|
|
| |
| from unsloth import FastVisionModel, get_chat_template |
| from unsloth.trainer import UnslothVisionDataCollator |
| from datasets import load_dataset |
| from trl import SFTTrainer, SFTConfig |
| from huggingface_hub import login |
|
|
| |
| token = os.environ.get("HF_TOKEN") |
| if token: |
| login(token=token) |
| logger.info("Logged in to Hugging Face Hub") |
| else: |
| logger.warning("HF_TOKEN not set - model upload may fail") |
|
|
| |
| print("\n[1/5] Loading model...") |
| start = time.time() |
|
|
| model, processor = FastVisionModel.from_pretrained( |
| args.base_model, |
| load_in_4bit=True, |
| use_gradient_checkpointing="unsloth", |
| ) |
|
|
| model = FastVisionModel.get_peft_model( |
| model, |
| finetune_vision_layers=True, |
| finetune_language_layers=True, |
| finetune_attention_modules=True, |
| finetune_mlp_modules=True, |
| r=args.lora_r, |
| lora_alpha=args.lora_alpha, |
| lora_dropout=0, |
| bias="none", |
| random_state=3407, |
| use_rslora=False, |
| loftq_config=None, |
| target_modules="all-linear", |
| ) |
|
|
| |
| processor = get_chat_template(processor, "gemma-3") |
| print(f"Model loaded in {time.time() - start:.1f}s") |
|
|
| |
| print("\n[2/5] Loading streaming dataset...") |
| start = time.time() |
|
|
| dataset = load_dataset( |
| args.dataset, |
| split="train", |
| streaming=True, |
| ) |
|
|
| |
| sample = next(iter(dataset)) |
| print(f"Dataset ready in {time.time() - start:.1f}s") |
| if "messages" in sample: |
| print(f" Sample has {len(sample['messages'])} messages") |
| if "images" in sample: |
| img_count = len(sample['images']) if isinstance(sample['images'], list) else 1 |
| print(f" Sample has {img_count} image(s)") |
|
|
| |
| dataset = load_dataset( |
| args.dataset, |
| split="train", |
| streaming=True, |
| ) |
|
|
| |
| print("\n[3/5] Configuring trainer...") |
|
|
| |
| FastVisionModel.for_training(model) |
|
|
| training_config = SFTConfig( |
| output_dir=args.save_local, |
| per_device_train_batch_size=args.batch_size, |
| gradient_accumulation_steps=args.gradient_accumulation, |
| gradient_checkpointing=True, |
| gradient_checkpointing_kwargs={"use_reentrant": False}, |
| max_grad_norm=0.3, |
| warmup_ratio=0.03, |
| max_steps=args.max_steps, |
| learning_rate=args.learning_rate, |
| logging_steps=max(1, args.max_steps // 20), |
| save_strategy="steps", |
| optim="adamw_torch_fused", |
| weight_decay=0.001, |
| lr_scheduler_type="cosine", |
| seed=3407, |
| |
| remove_unused_columns=False, |
| dataset_text_field="", |
| dataset_kwargs={"skip_prepare_dataset": True}, |
| max_length=args.max_seq_length, |
| |
| report_to="trackio", |
| run_name=f"vlm-streaming-{args.max_steps}steps", |
| ) |
|
|
| trainer = SFTTrainer( |
| model=model, |
| train_dataset=dataset, |
| processing_class=processor.tokenizer, |
| data_collator=UnslothVisionDataCollator(model, processor), |
| args=training_config, |
| ) |
|
|
| |
| print(f"\n[4/5] Training for {args.max_steps} steps...") |
| start = time.time() |
|
|
| trainer.train() |
|
|
| train_time = time.time() - start |
| print(f"\nTraining completed in {train_time / 60:.1f} minutes") |
| print(f" Speed: {args.max_steps / train_time:.2f} steps/s") |
|
|
| |
| print("\n[5/5] Saving model...") |
|
|
| |
| model.save_pretrained(args.save_local) |
| processor.save_pretrained(args.save_local) |
| print(f"Saved locally to {args.save_local}/") |
|
|
| |
| print(f"\nPushing to {args.output_repo}...") |
| model.push_to_hub(args.output_repo) |
| processor.push_to_hub(args.output_repo) |
| print(f"Model available at: https://huggingface.co/{args.output_repo}") |
|
|
| print("\n" + "=" * 70) |
| print("Done!") |
| print("=" * 70) |
|
|
|
|
| if __name__ == "__main__": |
| |
| if len(sys.argv) == 1: |
| print("=" * 70) |
| print("VLM Streaming Fine-tuning with Unsloth") |
| print("=" * 70) |
| print("\nFine-tune Vision-Language Models using streaming datasets.") |
| print("Data streams directly from the Hub - no disk space needed.") |
| print("\nFeatures:") |
| print(" - ~60% less VRAM with Unsloth optimizations") |
| print(" - 2x faster training vs standard methods") |
| print(" - Trackio integration for monitoring") |
| print(" - Works with any VLM dataset in conversation format") |
| print("\nExample usage:") |
| print("\n uv run vlm-streaming-sft-unsloth.py \\") |
| print(" --max-steps 500 \\") |
| print(" --output-repo your-username/vlm-finetuned") |
| print("\nHF Jobs example:") |
| print("\n hf jobs uv run vlm-streaming-sft-unsloth.py \\") |
| print(" --flavor a100-large \\") |
| print(" --secrets HF_TOKEN \\") |
| print(" -- \\") |
| print(" --max-steps 500 \\") |
| print(" --output-repo your-username/vlm-finetuned") |
| print("\nFor full help: uv run vlm-streaming-sft-unsloth.py --help") |
| print("=" * 70) |
| sys.exit(0) |
|
|
| main() |
|
|