-
Notifications
You must be signed in to change notification settings - Fork 354
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is it possible to use this library with sentence transformers? #548
Comments
I'll second this question. @ydennisy You say you're looking to generate quality embeddings. I'd imagine there are two fine-tuning tasks involved:
Also, have you thought about fine-tuning the underlying model instead? So, instead of taking a packaged sentence-transformer, taking the underlying model (e.g. BERT), fine-tuning that using MLM (masked language modelling), and only then using the sentence-transformer library to make that model into a sentence-transformer, which you then can train on labelled data. Haven't done this yet myself. |
Hey @leobaumgardt Yeah those are the steps I would like to take, that is why adapters seemed a good fit as there are tasks which can be done for each of these and then stacked. Hmm, not I have not explored that option yet - it does sound promising. However I would like to avoid porting models from HF to ST if at all possible. I find the ST library very simple to use but it is not overly flexible. Two ideas I am exploring:
Let me know if you have the same issues, would be happy to colab :) |
Hey @ydennisy, Also, we would be thankful if you could provide your solution once it works. So that others looking for the same problem can see how to solve it. |
This issue has been automatically marked as stale because it has been without activity for 90 days. This issue will be closed in 14 days unless you comment or remove the stale label. |
This issue was closed because it was stale for 14 days without any activity. |
Hi!
Amazing library and thanks for your work :)
I would like to use adapter with Sentence Transformers https://www.sbert.net/
I am able to load some of the pre-trained sbert models and use adapter to fine tune for tasks such as classification. However my downstream task is actually the production of quality embeddings for queries and text - so I would like to use the original multiple negatives ranking loss or triplet loss used in sbert.
Any help would be greatly appreciated!!!
The text was updated successfully, but these errors were encountered: