-
Notifications
You must be signed in to change notification settings - Fork 31k
handle inputs from Siglip/Siglip2 non-automapped encoder layers #41930
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Changes from 14 commits
ca68be8
76a14c7
4f93734
082dcf2
fe7c922
5aa7610
807983c
448dd63
91d34b0
40a9dc8
f54d0db
3ee3c56
c4400cb
92eeed6
b3db730
b7fec03
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -166,6 +166,10 @@ def test_flash_attn_2_inference_equivalence_right_padding(self): | |
| def test_sdpa_can_dispatch_on_flash(self): | ||
| pass | ||
|
|
||
| @unittest.skip(reason="This test is broken on A10 multi runners for now") | ||
| def test_multi_gpu_data_parallel_forward(self): | ||
| pass | ||
|
||
|
|
||
|
|
||
| class Siglip2VisionModelTester: | ||
| def __init__( | ||
|
|
||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Seems unrelated, no?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
you would believe that, right 👀
it wasn't AFAIK but it's related to when I shuffled around PreTrainedModel or not
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Dang wtf, no biggie but interesting