Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: add three new open clip roberta base models #860
feat: add three new open clip roberta base models #860
Changes from 13 commits
5524f46
06ec06f
7fcb813
c4beeca
5fbfb57
d5dd1ce
d937f13
fe2745b
7ddd51b
cce60df
1ef019d
a93d738
cc3514d
1e94a6e
0d2f871
5b0a65a
43b7a31
9efb369
848f0aa
b247228
b1fc4d6
5478afc
ba3aa44
2b7af82
06ad96a
8ef2090
8a36b8a
cf9595d
e43990b
1bf83ad
f918d65
d910ee8
423f788
04c130a
File filter
Filter by extension
Conversations
Jump to
There are no files selected for viewing
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I would suggest removing these asserts. It's much safe, but degrading the performance a bit.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
And what's more, from the function's parameter,
seq_len
. It seems that the flash-attention implementation can only be used for the text encoder. Is it can be applied to a vision transformer?There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes it could. Every image tensor first convert to sentence-like tensor before fed into model.