SFT-Qwen-SEA-LION-v4-8B-VL-OCT

LoRA adapter for Qwen-SEA-LION-v4-8B-VL fine-tuned on the OCT (Optical Coherence Tomography) modality from the OmniMedVQA dataset.

Training Details

  • Base model: aisingapore/Qwen-SEA-LION-v4-8B-VL
  • Method: SFT with LoRA (r=64, alpha=128)
  • Dataset: OmniMedVQA - OCT (Optical Coherence Tomography) modality
  • Epochs: 2
  • Learning rate: 1e-4 (cosine schedule)
  • Precision: bf16

Usage

from transformers import AutoProcessor, Qwen3VLForConditionalGeneration
from peft import PeftModel

base_model = Qwen3VLForConditionalGeneration.from_pretrained(
    "aisingapore/Qwen-SEA-LION-v4-8B-VL",
    torch_dtype="bfloat16",
    device_map="auto",
)
model = PeftModel.from_pretrained(base_model, "aagdeyogipramana/SFT-Qwen-SEA-LION-v4-8B-VL-OCT")
processor = AutoProcessor.from_pretrained("aagdeyogipramana/SFT-Qwen-SEA-LION-v4-8B-VL-OCT")

Citation

@article{lai2025med,
  title={Med-R1: Reinforcement Learning for Generalizable Medical Reasoning in Vision-Language Models},
  author={Lai, Yuxiang and Zhong, Jike and Li, Ming and Zhao, Shitian and Yang, Xiaofeng},
  journal={arXiv preprint arXiv:2503.13939},
  year={2025}
}
Downloads last month
3
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for aagdeyogipramana/SFT-Qwen-SEA-LION-v4-8B-VL-OCT

Adapter
(15)
this model

Collection including aagdeyogipramana/SFT-Qwen-SEA-LION-v4-8B-VL-OCT

Paper for aagdeyogipramana/SFT-Qwen-SEA-LION-v4-8B-VL-OCT