-
Notifications
You must be signed in to change notification settings - Fork 61
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Finetuning with lora output never ends. #29
Comments
same question |
Hi Both, Thanks for your interest in our work. I noticed you are using the wrong LLaMA3 base model that may have some issues with the tokenizer as reported in the earlier versions. I would recommend using the official meta-llama/Meta-Llama-3-8B as base version as they fixed the tokenizer issue which was effecting generation. Let me know if this solves the issue. Thanks and Good Luck |
Nice work! I am using the latest llava-llama3 model downloaded from huggingface and attempting to use it directly for Lora. When I directly use the model without Lora, I will repeatedly output the final text content on my task until the maximum length, and I suspect it may be related to EOS. In addition, when I try to use Lora, the output becomes strange and even produces some content that is not a word. Is this related to me directly using the original version of llava's finetune task_lora? I only replaced the llava-llama3 version with the dialogue template llama3 and the base model hf. Thank you again for your work. Very helpful to me :) |
same problem. The output from model is a bunch of exclamation marks |
have the same proble when i set model_bash=llava_meta_llama,but when set model_base=llama,the result is correct, how i finetuen from the llava_llama3 instead of from scrach the llama3? |
Hi, Thanks for your wonderful work.
I am struggling using my lora tuned model.
I conducted following steps
The text was updated successfully, but these errors were encountered: