-
Notifications
You must be signed in to change notification settings - Fork 34
add_text2video_ort_pipeline #105
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
add_text2video_ort_pipeline #105
Conversation
0b37b3d to
5d6c1f4
Compare
60325d7 to
86347be
Compare
86347be to
307c429
Compare
c8f37ec to
6840a70
Compare
6840a70 to
b19e46a
Compare
b19e46a to
1dc9970
Compare
| class VideoOnnxConfig(OnnxConfig): | ||
| """Handles video architectures.""" | ||
|
|
||
| DUMMY_INPUT_GENERATOR_CLASSES = (DummyVideoInputGenerator, DummyTimestepInputGenerator) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't think an abstract video onnx config is needed as it doesn't really abstract much here
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi ! thanks a lot for the contribution !
is the PR finished ? I believe there also needs to be a method that describes how the Wan pipeline is split and which components it needs to export / use. Also some testing with a tiny model on the exporters and onnxruntime side would be great.
|
@naomili0924 let's rather follow the same design we did with sana, i.e. having a specific function for splitting the wan pipelines. |
|
Relates to:
huggingface/diffusers#12846
huggingface/optimum#2389
This Pull Request is adding a uniform AutoText2VideoORTPipeline as requested from: huggingface/optimum#2168
Result:
output.mp4