| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| """ |
| Fine-tune LightOnOCR on OCR datasets. |
| |
| LightOnOCR is an end-to-end trainable vision-language model specifically designed for OCR tasks. |
| This script enables fine-tuning on custom datasets for improved performance on specific domains, |
| languages, or document types. |
| |
| Examples: |
| # Basic fine-tuning on IAM handwriting dataset |
| uv run lightonocr-finetune.py \ |
| --dataset-id HuggingFaceM4/FineVision \ |
| --subset iam \ |
| --output-dir ./lightonocr-iam \ |
| --epochs 2 |
| |
| # Fine-tune with frozen language model to save memory |
| uv run lightonocr-finetune.py \ |
| --dataset-id HuggingFaceM4/FineVision \ |
| --subset olmOCR-mix-0225-documents \ |
| --output-dir ./lightonocr-docs \ |
| --freeze-language \ |
| --batch-size 8 |
| |
| # Stream large datasets to reduce memory usage |
| uv run lightonocr-finetune.py \ |
| --dataset-id HuggingFaceM4/FineVision \ |
| --subset olmOCR-mix-0225-books \ |
| --output-dir ./lightonocr-books \ |
| --streaming \ |
| --shuffle-buffer-size 10000 \ |
| --max-train-samples 5000 # Will auto-calculate max-steps |
| |
| # Push to Hub with evaluation metrics |
| uv run lightonocr-finetune.py \ |
| --dataset-id HuggingFaceM4/FineVision \ |
| --subset iam \ |
| --hub-model-id username/lightonocr-iam \ |
| --push-to-hub \ |
| --eval-samples 100 |
| |
| # Run on HF Jobs with GPU and streaming |
| hf jobs run --gpu l4x1 \ |
| uv run lightonocr-finetune.py \ |
| --dataset-id custom/large-ocr-dataset \ |
| --output-dir ./custom-ocr \ |
| --streaming \ |
| --epochs 3 |
| """ |
|
|
| import argparse |
| import logging |
| import os |
| import sys |
| from datetime import datetime |
| from pathlib import Path |
| from typing import Dict, Optional |
|
|
| import torch |
| from datasets import load_dataset |
| from huggingface_hub import login |
| from jiwer import cer, wer |
| from tqdm import tqdm |
| from transformers import ( |
| AutoProcessor, |
| LightOnOCRForConditionalGeneration, |
| Trainer, |
| TrainingArguments, |
| ) |
|
|
| logging.basicConfig( |
| level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s" |
| ) |
| logger = logging.getLogger(__name__) |
|
|
| os.environ["HF_XET_HIGH_PERFORMANCE"] = "1" |
|
|
| |
| ASSISTANT_START_PATTERN = [151645, 1699, 151644, 77091, 1699] |
| DEFAULT_MAX_LENGTH = 1024 |
| DEFAULT_LONGEST_EDGE = 700 |
|
|
|
|
| class OCRDataCollator: |
| """Data collator for OCR fine-tuning.""" |
|
|
| def __init__( |
| self, |
| processor, |
| max_length=DEFAULT_MAX_LENGTH, |
| longest_edge=DEFAULT_LONGEST_EDGE, |
| ): |
| self.processor = processor |
| self.max_length = max_length |
| self.longest_edge = longest_edge |
|
|
| def __call__(self, examples): |
| batch_messages = [] |
| batch_images = [] |
|
|
| for example in examples: |
| example_images = example["images"] |
| example_texts = example["texts"] |
|
|
| |
| if len(example_images) != 1 or len(example_texts) != 1: |
| logger.warning( |
| f"Skipping example with {len(example_images)} images and {len(example_texts)} texts" |
| ) |
| continue |
|
|
| image = example_images[0].convert("RGB") |
| batch_images.append(image) |
|
|
| |
| conversation = example_texts[0] |
| assistant_text = conversation.get("assistant", "").strip() |
|
|
| messages = [ |
| {"role": "user", "content": [{"type": "image"}]}, |
| { |
| "role": "assistant", |
| "content": [{"type": "text", "text": assistant_text}], |
| }, |
| ] |
| batch_messages.append(messages) |
|
|
| if not batch_images: |
| return None |
|
|
| |
| texts = [ |
| self.processor.apply_chat_template( |
| messages, tokenize=False, add_generation_prompt=False |
| ) |
| for messages in batch_messages |
| ] |
|
|
| |
| inputs = self.processor( |
| text=texts, |
| images=batch_images, |
| return_tensors="pt", |
| padding=True, |
| truncation=True, |
| max_length=self.max_length, |
| size={"longest_edge": self.longest_edge}, |
| ) |
|
|
| |
| labels = inputs["input_ids"].clone() |
| pad_token_id = self.processor.tokenizer.pad_token_id |
|
|
| for i in range(len(labels)): |
| full_ids = inputs["input_ids"][i].tolist() |
|
|
| |
| assistant_content_start = None |
|
|
| |
| for idx in range(len(full_ids) - len(ASSISTANT_START_PATTERN)): |
| if ( |
| full_ids[idx : idx + len(ASSISTANT_START_PATTERN)] |
| == ASSISTANT_START_PATTERN |
| ): |
| assistant_content_start = idx + len(ASSISTANT_START_PATTERN) |
| break |
|
|
| if assistant_content_start is None: |
| |
| |
| labels[i, :] = -100 |
| else: |
| |
| labels[i, :] = -100 |
|
|
| |
| for idx in range(assistant_content_start, len(full_ids)): |
| if full_ids[idx] == pad_token_id: |
| break |
| labels[i, idx] = inputs["input_ids"][i, idx] |
|
|
| |
| labels[i, inputs["input_ids"][i] == pad_token_id] = -100 |
|
|
| inputs["labels"] = labels |
|
|
| |
| inputs["pixel_values"] = inputs["pixel_values"].to(torch.bfloat16) |
|
|
| return inputs |
|
|
|
|
| def evaluate_model( |
| model, |
| processor, |
| dataset, |
| num_samples: int = 50, |
| batch_size: int = 8, |
| device: str = "cuda", |
| description: str = "Model", |
| is_streaming: bool = False, |
| ) -> Dict[str, float]: |
| """ |
| Evaluate model on dataset and compute OCR metrics. |
| |
| Args: |
| model: The model to evaluate |
| processor: The processor for the model |
| dataset: Dataset to evaluate on (can be streaming or regular) |
| num_samples: Number of samples to evaluate |
| batch_size: Batch size for evaluation |
| device: Device to run evaluation on |
| description: Description for logging |
| is_streaming: Whether the dataset is a streaming dataset |
| |
| Returns: |
| Dictionary with CER, WER, and perfect match count |
| """ |
| model.eval() |
| predictions = [] |
| ground_truths = [] |
|
|
| logger.info(f"Evaluating {description} on {num_samples} samples...") |
|
|
| |
| if is_streaming: |
| |
| samples_processed = 0 |
| batch_samples = [] |
|
|
| for sample in tqdm(dataset, total=num_samples, desc="Evaluating"): |
| if samples_processed >= num_samples: |
| break |
|
|
| batch_samples.append(sample) |
| samples_processed += 1 |
|
|
| |
| if len(batch_samples) == batch_size or samples_processed == num_samples: |
| batch_images = [[s["images"][0]] for s in batch_samples] |
| batch_ground_truths = [ |
| s["texts"][0]["assistant"].strip() for s in batch_samples |
| ] |
|
|
| |
| messages = [{"role": "user", "content": [{"type": "image"}]}] |
| text = processor.apply_chat_template( |
| messages, tokenize=False, add_generation_prompt=True |
| ) |
| texts = [text] * len(batch_images) |
|
|
| inputs = processor( |
| text=texts, |
| images=batch_images, |
| return_tensors="pt", |
| padding=True, |
| truncation=True, |
| max_length=DEFAULT_MAX_LENGTH, |
| size={"longest_edge": DEFAULT_LONGEST_EDGE}, |
| ).to(device) |
| inputs["pixel_values"] = inputs["pixel_values"].to(torch.bfloat16) |
|
|
| |
| with torch.no_grad(): |
| outputs = model.generate( |
| **inputs, max_new_tokens=512, do_sample=True |
| ) |
|
|
| input_length = inputs["input_ids"].shape[1] |
| generated_ids = outputs[:, input_length:] |
| batch_predictions = processor.batch_decode( |
| generated_ids, skip_special_tokens=True |
| ) |
| batch_predictions = [p.strip() for p in batch_predictions] |
|
|
| predictions.extend(batch_predictions) |
| ground_truths.extend(batch_ground_truths) |
| batch_samples = [] |
| else: |
| |
| for start_idx in tqdm(range(0, min(num_samples, len(dataset)), batch_size)): |
| end_idx = min(start_idx + batch_size, num_samples, len(dataset)) |
| batch_samples = [dataset[i] for i in range(start_idx, end_idx)] |
|
|
| batch_images = [[s["images"][0]] for s in batch_samples] |
| batch_ground_truths = [ |
| s["texts"][0]["assistant"].strip() for s in batch_samples |
| ] |
|
|
| |
| messages = [{"role": "user", "content": [{"type": "image"}]}] |
| text = processor.apply_chat_template( |
| messages, tokenize=False, add_generation_prompt=True |
| ) |
| texts = [text] * len(batch_images) |
|
|
| inputs = processor( |
| text=texts, |
| images=batch_images, |
| return_tensors="pt", |
| padding=True, |
| truncation=True, |
| max_length=DEFAULT_MAX_LENGTH, |
| size={"longest_edge": DEFAULT_LONGEST_EDGE}, |
| ).to(device) |
| inputs["pixel_values"] = inputs["pixel_values"].to(torch.bfloat16) |
|
|
| |
| with torch.no_grad(): |
| outputs = model.generate(**inputs, max_new_tokens=512, do_sample=True) |
|
|
| input_length = inputs["input_ids"].shape[1] |
| generated_ids = outputs[:, input_length:] |
| batch_predictions = processor.batch_decode( |
| generated_ids, skip_special_tokens=True |
| ) |
| batch_predictions = [p.strip() for p in batch_predictions] |
|
|
| predictions.extend(batch_predictions) |
| ground_truths.extend(batch_ground_truths) |
|
|
| |
| cer_score = cer(ground_truths, predictions) * 100 |
| wer_score = wer(ground_truths, predictions) * 100 |
| perfect_matches = sum( |
| 1 for pred, gt in zip(predictions, ground_truths) if pred == gt |
| ) |
|
|
| actual_samples = len(predictions) |
| logger.info( |
| f"CER: {cer_score:.2f}% | WER: {wer_score:.2f}% | Perfect: {perfect_matches}/{actual_samples}" |
| ) |
|
|
| |
| for i in range(min(3, len(predictions))): |
| match = "✅" if predictions[i] == ground_truths[i] else "❌" |
| logger.info( |
| f"{match} Sample {i + 1}: '{predictions[i][:50]}...' vs '{ground_truths[i][:50]}...'" |
| ) |
|
|
| return { |
| "cer": cer_score, |
| "wer": wer_score, |
| "perfect_matches": perfect_matches, |
| "total_samples": actual_samples, |
| } |
|
|
|
|
| def create_model_card_content( |
| model_id: str, |
| dataset_id: str, |
| subset: Optional[str], |
| base_metrics: Dict[str, float], |
| finetuned_metrics: Dict[str, float], |
| training_args: TrainingArguments, |
| freeze_config: Dict[str, bool], |
| ) -> str: |
| """Generate model card content with training details and metrics.""" |
|
|
| |
| cer_improvement = base_metrics["cer"] - finetuned_metrics["cer"] |
| wer_improvement = base_metrics["wer"] - finetuned_metrics["wer"] |
| perfect_improvement = ( |
| finetuned_metrics["perfect_matches"] - base_metrics["perfect_matches"] |
| ) |
|
|
| |
| frozen_components = [comp for comp, is_frozen in freeze_config.items() if is_frozen] |
| frozen_str = ( |
| ", ".join(frozen_components) if frozen_components else "None (full fine-tuning)" |
| ) |
|
|
| dataset_str = f"{dataset_id}/{subset}" if subset else dataset_id |
|
|
| content = f"""--- |
| license: mit |
| tags: |
| - vision |
| - ocr |
| - document-understanding |
| - transformers |
| base_model: lightonai/LightOnOCR-1B-1025 |
| datasets: |
| - {dataset_id} |
| metrics: |
| - cer |
| - wer |
| library_name: transformers |
| --- |
| |
| # {model_id.split("/")[-1]} |
| |
| This model is a fine-tuned version of [LightOnOCR-1B-1025](https://huggingface.co/lightonai/LightOnOCR-1B-1025) on the {dataset_str} dataset. |
| |
| ## Model Description |
| |
| LightOnOCR is an end-to-end trainable vision-language model specifically designed for OCR tasks. This fine-tuned version has been optimized for improved performance on the target dataset. |
| |
| ## Training Details |
| |
| ### Dataset |
| - **Source**: {dataset_str} |
| - **Training samples**: {training_args.num_train_epochs * (training_args.per_device_train_batch_size * training_args.gradient_accumulation_steps)} |
| - **Validation samples**: Used for model selection |
| |
| ### Training Configuration |
| - **Epochs**: {training_args.num_train_epochs} |
| - **Batch size**: {training_args.per_device_train_batch_size} (effective: {training_args.per_device_train_batch_size * training_args.gradient_accumulation_steps}) |
| - **Learning rate**: {training_args.learning_rate} |
| - **Frozen components**: {frozen_str} |
| - **Hardware**: GPU with mixed precision (bf16) |
| |
| ## Evaluation Results |
| |
| ### Performance Comparison |
| |
| | Metric | Base Model | Fine-tuned | Improvement | |
| |--------|------------|------------|-------------| |
| | **CER (%)** | {base_metrics["cer"]:.2f} | {finetuned_metrics["cer"]:.2f} | {cer_improvement:+.2f} | |
| | **WER (%)** | {base_metrics["wer"]:.2f} | {finetuned_metrics["wer"]:.2f} | {wer_improvement:+.2f} | |
| | **Perfect Matches** | {base_metrics["perfect_matches"]}/{base_metrics["total_samples"]} | {finetuned_metrics["perfect_matches"]}/{finetuned_metrics["total_samples"]} | {perfect_improvement:+d} | |
| |
| *Lower is better for CER and WER. Evaluation performed on {finetuned_metrics["total_samples"]} test samples.* |
| |
| ## Usage |
| |
| ```python |
| from transformers import AutoProcessor, LightOnOCRForConditionalGeneration |
| from PIL import Image |
| import torch |
| |
| # Load model and processor |
| model = LightOnOCRForConditionalGeneration.from_pretrained( |
| "{model_id}", |
| torch_dtype=torch.bfloat16, |
| device_map="auto" |
| ) |
| processor = AutoProcessor.from_pretrained("{model_id}") |
| |
| # Prepare image |
| image = Image.open("your_image.jpg").convert("RGB") |
| |
| # Create prompt |
| messages = [ |
| {{"role": "user", "content": [{{"type": "image"}}]}} |
| ] |
| |
| # Process and generate |
| text = processor.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) |
| inputs = processor( |
| text=[text], |
| images=[[image]], |
| return_tensors="pt", |
| max_length=1024 |
| ).to(model.device) |
| |
| outputs = model.generate(**inputs, max_new_tokens=512) |
| generated_text = processor.decode(outputs[0][inputs['input_ids'].shape[1]:], skip_special_tokens=True) |
| print(generated_text) |
| ``` |
| |
| ## Training Script |
| |
| This model was trained using the UV Scripts training pipeline. To reproduce or further fine-tune: |
| |
| ```bash |
| uv run https://huggingface.co/datasets/uv-scripts/transformers-training/raw/main/lightonocr-finetune.py \\ |
| --dataset-id {dataset_id} \\ |
| {"--subset " + subset if subset else ""} \\ |
| --output-dir ./model \\ |
| --epochs {training_args.num_train_epochs} |
| ``` |
| |
| ## Citation |
| |
| If you use this model, please cite: |
| |
| ```bibtex |
| @misc{{lightonocr2024, |
| title={{LightOnOCR: End-to-End Trainable OCR Model}}, |
| author={{LightOn AI}}, |
| year={{2024}}, |
| url={{https://huggingface.co/blog/lightonai/lightonocr}} |
| }} |
| ``` |
| |
| ## License |
| |
| This model is released under the MIT license. |
| |
| --- |
| |
| *Generated on {datetime.now().strftime("%Y-%m-%d")} using [UV Scripts](https://huggingface.co/uv-scripts)* |
| """ |
|
|
| return content |
|
|
|
|
| def main(): |
| parser = argparse.ArgumentParser( |
| description="Fine-tune LightOnOCR on OCR datasets", |
| formatter_class=argparse.RawDescriptionHelpFormatter, |
| ) |
|
|
| |
| parser.add_argument( |
| "--dataset-id", |
| type=str, |
| default="HuggingFaceM4/FineVision", |
| help="HuggingFace dataset ID", |
| ) |
| parser.add_argument( |
| "--subset", |
| type=str, |
| default="iam", |
| choices=["iam", "olmOCR-mix-0225-books", "olmOCR-mix-0225-documents"], |
| help="Dataset subset to use (for FineVision)", |
| ) |
| parser.add_argument( |
| "--train-split", |
| type=str, |
| default="train[:85%]", |
| help="Training split specification", |
| ) |
| parser.add_argument( |
| "--val-split", |
| type=str, |
| default="train[85%:95%]", |
| help="Validation split specification", |
| ) |
| parser.add_argument( |
| "--test-split", type=str, default="train[95%:]", help="Test split specification" |
| ) |
|
|
| |
| parser.add_argument( |
| "--streaming", |
| action="store_true", |
| help="Use dataset streaming to reduce memory usage (Note: uses full training set, ignores train-split percentages)", |
| ) |
| parser.add_argument( |
| "--shuffle-buffer-size", |
| type=int, |
| default=10000, |
| help="Buffer size for shuffling when using streaming (default: 10000)", |
| ) |
| parser.add_argument( |
| "--max-train-samples", |
| type=int, |
| help="Maximum number of training samples when streaming (useful for quick experiments)", |
| ) |
|
|
| |
| parser.add_argument( |
| "--model-id", |
| type=str, |
| default="lightonai/LightOnOCR-1B-1025", |
| help="Base model ID", |
| ) |
| parser.add_argument( |
| "--freeze-vision", action="store_true", help="Freeze vision encoder" |
| ) |
| parser.add_argument( |
| "--freeze-language", action="store_true", help="Freeze language model" |
| ) |
| parser.add_argument( |
| "--freeze-projection", |
| action="store_true", |
| help="Freeze vision projection layer", |
| ) |
|
|
| |
| parser.add_argument( |
| "--output-dir", type=str, required=True, help="Directory to save the model" |
| ) |
| parser.add_argument( |
| "--epochs", type=int, default=2, help="Number of training epochs" |
| ) |
| parser.add_argument( |
| "--batch-size", type=int, default=4, help="Training batch size per device" |
| ) |
| parser.add_argument( |
| "--gradient-accumulation", |
| type=int, |
| default=4, |
| help="Gradient accumulation steps", |
| ) |
| parser.add_argument( |
| "--learning-rate", type=float, default=6e-5, help="Learning rate" |
| ) |
| parser.add_argument( |
| "--warmup-steps", type=int, default=10, help="Number of warmup steps" |
| ) |
| parser.add_argument( |
| "--eval-steps", type=int, default=50, help="Evaluation interval (in steps)" |
| ) |
| parser.add_argument( |
| "--save-steps", |
| type=int, |
| default=500, |
| help="Save checkpoint interval (in steps)", |
| ) |
| parser.add_argument( |
| "--max-length", type=int, default=1024, help="Maximum sequence length" |
| ) |
| parser.add_argument( |
| "--longest-edge", type=int, default=700, help="Longest edge for image resizing" |
| ) |
|
|
| |
| parser.add_argument( |
| "--eval-samples", type=int, default=100, help="Number of samples for evaluation" |
| ) |
| parser.add_argument( |
| "--eval-batch-size", type=int, default=8, help="Batch size for evaluation" |
| ) |
| parser.add_argument( |
| "--skip-base-eval", action="store_true", help="Skip base model evaluation" |
| ) |
|
|
| |
| parser.add_argument( |
| "--hub-model-id", type=str, help="HuggingFace Hub model ID for pushing" |
| ) |
| parser.add_argument( |
| "--push-to-hub", action="store_true", help="Push model to HuggingFace Hub" |
| ) |
| parser.add_argument("--hf-token", type=str, help="HuggingFace API token") |
| parser.add_argument( |
| "--private", action="store_true", help="Make the model private on Hub" |
| ) |
|
|
| |
| parser.add_argument( |
| "--max-samples", type=int, help="Limit number of training samples (for testing)" |
| ) |
| parser.add_argument("--seed", type=int, default=42, help="Random seed") |
| parser.add_argument( |
| "--max-steps", |
| type=int, |
| help="Maximum number of training steps (auto-calculated for streaming if not specified)" |
| ) |
|
|
| args = parser.parse_args() |
|
|
| |
| if not torch.cuda.is_available(): |
| logger.error("CUDA is not available. This script requires a GPU.") |
| logger.info("To run on HF Jobs with GPU:") |
| logger.info( |
| f"hf jobs run --gpu l4x1 uv run {__file__} --dataset-id {args.dataset_id} --output-dir {args.output_dir}" |
| ) |
| sys.exit(1) |
|
|
| device = "cuda" |
| logger.info(f"Using GPU: {torch.cuda.get_device_name(0)}") |
|
|
| |
| os.environ["PYTORCH_CUDA_ALLOC_CONF"] = "expandable_segments:True" |
| os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1" |
| torch.set_float32_matmul_precision("high") |
|
|
| |
| if args.push_to_hub: |
| token = args.hf_token or os.environ.get("HF_TOKEN") |
| if token: |
| login(token=token) |
| else: |
| logger.error("HF_TOKEN required for push_to_hub") |
| sys.exit(1) |
|
|
| |
| logger.info(f"Loading dataset: {args.dataset_id}/{args.subset}") |
|
|
| if args.streaming: |
| logger.info("Using streaming mode for dataset loading") |
| |
| |
| train_ds = load_dataset( |
| args.dataset_id, args.subset, split="train", streaming=True |
| ) |
|
|
| |
| |
| |
| full_ds = load_dataset(args.dataset_id, args.subset, split="train") |
| total_size = len(full_ds) |
|
|
| |
| train_end = int(0.85 * total_size) |
| val_end = int(0.95 * total_size) |
|
|
| |
| val_ds = full_ds.select(range(train_end, val_end)) |
| test_ds = full_ds.select(range(val_end, total_size)) |
|
|
| |
| del full_ds |
|
|
| |
| train_ds = train_ds.shuffle( |
| seed=args.seed, buffer_size=args.shuffle_buffer_size |
| ) |
|
|
| |
| if args.max_samples or args.max_train_samples: |
| max_samples = args.max_samples or args.max_train_samples |
| train_ds = train_ds.take(max_samples) |
| logger.info(f"Limited training to {max_samples} samples (streaming mode)") |
|
|
| logger.info( |
| f"Dataset loaded - Training: streaming (full train set), Val: {len(val_ds)}, Test: {len(test_ds)}" |
| ) |
| logger.info( |
| "Note: When streaming, using full training set. Use --max-train-samples to limit." |
| ) |
| else: |
| |
| train_ds = load_dataset(args.dataset_id, args.subset, split=args.train_split) |
| val_ds = load_dataset(args.dataset_id, args.subset, split=args.val_split) |
| test_ds = load_dataset(args.dataset_id, args.subset, split=args.test_split) |
|
|
| |
| if args.max_samples: |
| train_ds = train_ds.select(range(min(args.max_samples, len(train_ds)))) |
| logger.info(f"Limited training to {len(train_ds)} samples") |
|
|
| logger.info( |
| f"Dataset sizes - Train: {len(train_ds)}, Val: {len(val_ds)}, Test: {len(test_ds)}" |
| ) |
|
|
| |
| logger.info(f"Loading processor from {args.model_id}") |
| processor = AutoProcessor.from_pretrained(args.model_id) |
| processor.tokenizer.padding_side = "left" |
|
|
| |
| logger.info(f"Loading model from {args.model_id}") |
| model = LightOnOCRForConditionalGeneration.from_pretrained( |
| args.model_id, |
| torch_dtype=torch.bfloat16, |
| attn_implementation="sdpa", |
| device_map="auto", |
| ).to(device) |
|
|
| |
| freeze_config = { |
| "vision_encoder": args.freeze_vision, |
| "language_model": args.freeze_language, |
| "vision_projection": args.freeze_projection, |
| } |
|
|
| if args.freeze_vision: |
| for param in model.model.vision_encoder.parameters(): |
| param.requires_grad = False |
| logger.info("Vision encoder frozen") |
|
|
| if args.freeze_language: |
| for param in model.model.language_model.parameters(): |
| param.requires_grad = False |
| logger.info("Language model frozen") |
|
|
| if args.freeze_projection: |
| for param in model.model.vision_projection.parameters(): |
| param.requires_grad = False |
| logger.info("Vision projection frozen") |
|
|
| |
| total_params = sum(p.numel() for p in model.parameters()) |
| trainable_params = sum(p.numel() for p in model.parameters() if p.requires_grad) |
| logger.info(f"Total parameters: {total_params:,}") |
| logger.info( |
| f"Trainable parameters: {trainable_params:,} ({100 * trainable_params / total_params:.2f}%)" |
| ) |
|
|
| |
| base_metrics = { |
| "cer": 0.0, |
| "wer": 0.0, |
| "perfect_matches": 0, |
| "total_samples": args.eval_samples, |
| } |
| if not args.skip_base_eval: |
| logger.info("\n" + "=" * 80) |
| logger.info("EVALUATING BASE MODEL") |
| logger.info("=" * 80) |
| base_metrics = evaluate_model( |
| model, |
| processor, |
| test_ds, |
| num_samples=args.eval_samples, |
| batch_size=args.eval_batch_size, |
| device=device, |
| description="Base model", |
| is_streaming=False, |
| ) |
| torch.cuda.empty_cache() |
|
|
| |
| data_collator = OCRDataCollator( |
| processor, max_length=args.max_length, longest_edge=args.longest_edge |
| ) |
|
|
| |
| max_steps = None |
| if args.streaming: |
| if args.max_steps: |
| max_steps = args.max_steps |
| logger.info(f"Using user-specified max_steps: {max_steps}") |
| else: |
| |
| if args.max_train_samples: |
| |
| steps_per_epoch = args.max_train_samples // (args.batch_size * args.gradient_accumulation) |
| max_steps = steps_per_epoch * args.epochs |
| logger.info(f"Calculated max_steps from max_train_samples: {max_steps}") |
| else: |
| |
| |
| |
| max_steps = 1000 * args.epochs |
| logger.warning( |
| f"Streaming mode: Using default max_steps={max_steps}. " |
| f"Consider setting --max-steps or --max-train-samples for precise control." |
| ) |
|
|
| |
| |
| training_args_dict = { |
| "output_dir": args.output_dir, |
| "per_device_train_batch_size": args.batch_size, |
| "per_device_eval_batch_size": args.eval_batch_size, |
| "gradient_accumulation_steps": args.gradient_accumulation, |
| "learning_rate": args.learning_rate, |
| "weight_decay": 0.0, |
| "logging_steps": 50, |
| "eval_strategy": "steps", |
| "eval_steps": args.eval_steps, |
| "save_strategy": "steps", |
| "save_steps": args.save_steps, |
| "save_total_limit": 2, |
| "load_best_model_at_end": True, |
| "metric_for_best_model": "eval_loss", |
| "bf16": True, |
| "fp16": False, |
| "remove_unused_columns": False, |
| "dataloader_pin_memory": False, |
| "gradient_checkpointing": True, |
| "optim": "adamw_torch_fused" if torch.cuda.is_available() else "adamw_torch", |
| "warmup_steps": args.warmup_steps, |
| "lr_scheduler_type": "linear", |
| "push_to_hub": args.push_to_hub, |
| "hub_model_id": args.hub_model_id, |
| "hub_private_repo": args.private, |
| } |
| |
| |
| if args.streaming: |
| training_args_dict["max_steps"] = max_steps |
| |
| training_args_dict["num_train_epochs"] = args.epochs |
| else: |
| training_args_dict["num_train_epochs"] = args.epochs |
| |
| training_args = TrainingArguments(**training_args_dict) |
|
|
| |
| val_ds_small = val_ds.select(range(min(100, len(val_ds)))) |
|
|
| |
| trainer = Trainer( |
| model=model, |
| args=training_args, |
| train_dataset=train_ds, |
| eval_dataset=val_ds_small, |
| data_collator=data_collator, |
| ) |
|
|
| |
| logger.info("\n" + "=" * 80) |
| logger.info("STARTING TRAINING") |
| logger.info("=" * 80) |
| if args.streaming: |
| logger.info( |
| f"Training samples: streaming mode (max: {args.max_train_samples or 'unlimited'})" |
| ) |
| logger.info(f"Max steps: {max_steps}") |
| else: |
| logger.info(f"Training samples: {len(train_ds)}") |
| logger.info(f"Validation samples: {len(val_ds_small)}") |
| logger.info(f"Effective batch size: {args.batch_size * args.gradient_accumulation}") |
|
|
| trainer.train() |
|
|
| |
| logger.info("Saving model and processor...") |
| trainer.save_model(args.output_dir) |
| processor.save_pretrained(args.output_dir) |
|
|
| |
| logger.info("\n" + "=" * 80) |
| logger.info("EVALUATING FINE-TUNED MODEL") |
| logger.info("=" * 80) |
| finetuned_metrics = evaluate_model( |
| model, |
| processor, |
| test_ds, |
| num_samples=args.eval_samples, |
| batch_size=args.eval_batch_size, |
| device=device, |
| description="Fine-tuned model", |
| is_streaming=False, |
| ) |
|
|
| |
| if not args.skip_base_eval: |
| logger.info("\n" + "=" * 80) |
| logger.info("PERFORMANCE COMPARISON") |
| logger.info("=" * 80) |
| logger.info( |
| f"{'Metric':<20} {'Base':<12} {'Fine-tuned':<12} {'Improvement':<12}" |
| ) |
| logger.info("-" * 56) |
| logger.info( |
| f"{'CER (%)':<20} {base_metrics['cer']:<12.2f} {finetuned_metrics['cer']:<12.2f} {base_metrics['cer'] - finetuned_metrics['cer']:+.2f}" |
| ) |
| logger.info( |
| f"{'WER (%)':<20} {base_metrics['wer']:<12.2f} {finetuned_metrics['wer']:<12.2f} {base_metrics['wer'] - finetuned_metrics['wer']:+.2f}" |
| ) |
| logger.info( |
| f"{'Perfect Matches':<20} {base_metrics['perfect_matches']:<12} {finetuned_metrics['perfect_matches']:<12} {finetuned_metrics['perfect_matches'] - base_metrics['perfect_matches']:+d}" |
| ) |
| logger.info("=" * 80) |
|
|
| |
| if args.hub_model_id or args.push_to_hub: |
| model_id = args.hub_model_id or f"{args.output_dir.split('/')[-1]}" |
| logger.info("Creating model card with metrics...") |
|
|
| model_card_content = create_model_card_content( |
| model_id=model_id, |
| dataset_id=args.dataset_id, |
| subset=args.subset, |
| base_metrics=base_metrics, |
| finetuned_metrics=finetuned_metrics, |
| training_args=training_args, |
| freeze_config=freeze_config, |
| ) |
|
|
| |
| model_card_path = Path(args.output_dir) / "README.md" |
| model_card_path.write_text(model_card_content) |
| logger.info(f"Model card saved to {model_card_path}") |
|
|
| if args.push_to_hub: |
| logger.info(f"Pushing model to Hub: {args.hub_model_id}") |
| trainer.push_to_hub() |
| logger.info( |
| f"✅ Model available at: https://huggingface.co/{args.hub_model_id}" |
| ) |
|
|
| logger.info("\n✅ Training complete!") |
| logger.info(f"Model saved to: {args.output_dir}") |
|
|
| |
| logger.info("\n" + "=" * 80) |
| logger.info("To use the fine-tuned model:") |
| logger.info("=" * 80) |
| logger.info(f""" |
| from transformers import AutoProcessor, LightOnOCRForConditionalGeneration |
| from PIL import Image |
| |
| model = LightOnOCRForConditionalGeneration.from_pretrained("{args.output_dir}") |
| processor = AutoProcessor.from_pretrained("{args.output_dir}") |
| # ... rest of inference code |
| """) |
|
|
|
|
| if __name__ == "__main__": |
| if len(sys.argv) == 1: |
| print("LightOnOCR Fine-tuning Script\n") |
| print("Examples:") |
| print(" # Basic fine-tuning:") |
| print( |
| " uv run lightonocr-finetune.py --dataset-id HuggingFaceM4/FineVision --subset iam --output-dir ./model\n" |
| ) |
| print(" # With frozen components:") |
| print |
| " uv run lightonocr-finetune.py --freeze-language --output-dir ./model\n" |
| ) |
| print(" # Stream large datasets (memory-efficient):") |
| print( |
| " uv run lightonocr-finetune.py --streaming --shuffle-buffer-size 10000 --output-dir ./model\n" |
| ) |
| print(" # Push to Hub:") |
| print( |
| " uv run lightonocr-finetune.py --hub-model-id username/model --push-to-hub\n" |
| ) |
| print(" # Run on HF Jobs:") |
| print( |
| " hf jobs run --gpu l4x1 uv run lightonocr-finetune.py --streaming --output-dir ./model" |
| ) |
| sys.exit(0) |
|
|
| main() |
|
|