"direct fine tuned" model
#10
by bdytx5 - opened
Discussion
bdytx5
2 days ago
Could you release the base fine tune model without the CoT training? I am writing an article on this. Thanks
Xkev
Owner
2 days ago
•
edited 2 days ago
Hi, the base model is Llama-3.2-11B-Vision-Instruct: https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct
Xkev changed discussion status to closed 2 days ago
bdytx5
2 days ago
Sorry, I mean the "Direct Training" model on the LLaVA 100k dataset (eg just trained on the answers rather than the entire CoT data). from the paper "Here, LLaVA-o1 (with Direct Training) refers to the model trained directly on the original VQA dataset’s Q&A pairs"
Leave a comment
I can upload this model, but is there any specific usage of that model?
Xkev changed discussion status to closed
To benchmark against llava o1.