"direct fine tuned" model

#10
by bdytx5 - opened

Discussion

bdytx5
2 days ago

Could you release the base fine tune model without the CoT training? I am writing an article on this. Thanks

Xkev
Owner
2 days ago

edited 2 days ago

Hi, the base model is Llama-3.2-11B-Vision-Instruct: https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct

Xkev changed discussion status to closed 2 days ago

bdytx5
2 days ago

Sorry, I mean the "Direct Training" model on the LLaVA 100k dataset (eg just trained on the answers rather than the entire CoT data). from the paper "Here, LLaVA-o1 (with Direct Training) refers to the model trained directly on the original VQA dataset’s Q&A pairs"

Leave a comment

I can upload this model, but is there any specific usage of that model?

Xkev changed discussion status to closed

To benchmark against llava o1.

Sign up or log in to comment