-
Notifications
You must be signed in to change notification settings - Fork 202
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can't we train and fine tune the Llavamed model #87
Comments
I fine tuned the model with the code here https://github.com/haotian-liu/LLaVA/blob/main/scripts/v1_5/finetune_task_lora.sh |
@thedaffodil can you explain a bit more? do you mean you cloned the original llava repo and run the training specified in the link you provided but changed the weights to be the weights of llava-med? |
@thedaffodil Is it to modify the weight and directly fine tune it |
@thedaffodil Is the environment Llavamed or Llava |
name: llava
my yaml file is like above. I use llava repo with llava-med weights |
i am trying to finetune llava initialized with llava-med on my own task. so far i tried running llava/train/train_mem.py but i notice that the loading of the model is done with LLavaLLama model instead of a mistral one, and i can't figure out where to modify this. any ideas? and generally where can i find more info on how to finetune llava med? |
I use the this command: deepspeed llava/train/train_mem.py after that I merged the output model and the base model to get weights with the code in the link below then I could use fine tuned model to eval. you can ask your further questions via my email if you need help. |
@thedaffodil I am here https://huggingface.co/microsoft/llava-med-v1.5-mistral-7b/tree/main Download the model from above and fine tune it with your script. The result shows that you are using a model of type llava_stistral to instruct a model of type llava_1lama This is not supported for all configurations of models and can yield errors. llava/train/train_mem.py FAILED Failures: Root Cause (first observed failure): |
while you are fine-tuning, your output model folder name should consist "finetune" not "llava". while you are merging, your output folder name should consist "llava" |
No description provided.
The text was updated successfully, but these errors were encountered: