You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Y is the generation speed so slow with sdxl I'm trying to use playground 2.5 but it takes forever to get a pic then it gives me a black image so I waited all that time for norhing,will you be making better optimizations soon BTW there is no support apk for 8gen1 it says 8gen1+ I tried using it and it just crashes when loading quantization model.
The text was updated successfully, but these errors were encountered:
There's nothing I can do about the speed — it’s not related to the Flutter/Dart side. It has to do with the ggml library, and the inference is slow on mobile due to the massive compute workload. That’s also why you don’t see many local diffusion inference apps around.
As for CPU performance, ggml is already fairly well optimized. I might revisit KleidAI microkernels, but in my last tests, they didn’t perform well with more than 4 threads.
The best path forward is to use distilled models that can converge in a single step. I’ve already added some: leejet/stable-diffusion.cpp#675
and plan to include more in the future.
I tested the app on a Snapdragon 8 Gen 3 and it works fine. Some vendors may not support custom CPU instructions, so I recommend sticking with the generic APK. On this device, I get around 8s/it for SD1.5, which is decent — you can get good results in under a minute.
The dark image output issue is due to SDXL’s VAE having NaN issues in FP16. As mentioned in a previous issue, Playground 2.5 also needs a custom scheduler anyway.
For now, I recommend sticking to distilled models with CFG-free sampling
Y is the generation speed so slow with sdxl I'm trying to use playground 2.5 but it takes forever to get a pic then it gives me a black image so I waited all that time for norhing,will you be making better optimizations soon BTW there is no support apk for 8gen1 it says 8gen1+ I tried using it and it just crashes when loading quantization model.
The text was updated successfully, but these errors were encountered: