-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
how to infer text-img pair demo? #8
Comments
We will update readme about this today. Will let you know after we add this. Thanks for you reminding. |
Same question here. btw, do you have a plan to release CC Finetuned LLM? |
We’ll do our best to release it within 24 hours. Thank you for the reminder. If you have any other requests, feel free to let us know. We’re happy to release whatever we can, as long as it complies with safety regulations. |
Any updates on this @Yif-Yang? |
@eek |
really thanks for your understanding, we will look at it tonight.😁 |
@BIGBALLOBA You are really kind and thoughtful🌹🌹🌹 |
Didn't want to come in rude, you guys @Yif-Yang did a fantastic job! I'm really curious and excited to play with it, so looking forward to it 😀 I'll wait patiently 😀 Congrats again! |
@eek @BIGBALLON @WinstonDeng @mtodd We have updated the caption contrastive fine-tuned version of Llama3-8B-CC (https://huggingface.co/microsoft/LLM2CLIP-Llama-3-8B-Instruct-CC-Finetuned) to assist with your retrieval experiments and training of your own CLIP models. Additionally, the parameters for our adapter and projector have been made available in our OpenAI ViT-L repository (https://huggingface.co/microsoft/LLM2CLIP-Openai-L-14-336). The retrieval testing methods are documented in the model card for reference. Our tests show retrieval performance exceeding the results reported in the paper, and we encourage you to try it out. Regarding the EVA series of models, there have been precision mismatches during the conversion to Hugging Face, which are currently being fixed. Updates will be released progressively. Furthermore, we will provide detailed instructions on how to use LLM2CLIP to fine-tune your own CLIP models in about a week—please stay tuned! |
expect the EVA series of models |
How about the text encoder for EVA series? |
will release soon, maybe today |
Thank you very much for your work! May I ask when the data preprocessing will be released? |
Should be around today or tomorrow I think. |
@konioy @JENNSHIUAN We just updated EVA02's pytorch ckpt. We will try update safe tensor version tomorrow. |
Great work, thanks a lot @Yif-Yang ! For the EVA02 model how does one encode the text? It doesn't seem to have the same |
will upload asap |
@vilhub We have updated the README in HuggingFace, which now includes usage examples. |
Using openai official text model, text embedding dim is 768, mismatching with llm2clip img embedding dim 1280.
The text was updated successfully, but these errors were encountered: