-
Notifications
You must be signed in to change notification settings - Fork 2.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Doesn't GISTEmbedLoss support DDP or DP? #2772
Comments
Hello! Thank you for reporting this - this is a bug caused by the retokenization that is required when your guide model has a different tokenizer than your training model: sentence-transformers/sentence_transformers/losses/CachedGISTEmbedLoss.py Lines 174 to 181 in e5c15a5
I think we can fix this by setting a sentence-transformers/sentence_transformers/losses/CachedGISTEmbedLoss.py Lines 135 to 137 in e5c15a5
Because here we do have the "normal" model rather than the DDP-wrapped model, so we can still access the tokenizer. We can then use the Until then, you can perhaps try to use a guide model with the same tokenizer as the model that you're training?
|
Thanks for reporting this. You can use (Cached)GISTEmbedLoss again with DDP/DP by installing the "bleeding edge" version of sentence-transformers:
|
But it doesn't work... Error Message
I ran "pip install git+https://github.com/UKPLab/sentence-transformers.git" in my virtual environment. Therefore, I experimented with changing to the version below.
|
When running CachedGISTEmbedLoss with DDP via the torchrun command, the following error occurs. The same error occurs even when running with DP. I built the anaconda environment using the "pip install ." command from the github here. Below is the result of the pip list command. Can you tell me what to modify?
Sorry for bothering you every time. I need your help.
error log
pip list
The text was updated successfully, but these errors were encountered: