-
Hi all. I have an Acer A515-52G-78HE and it's have a GEFORCE MX 130. I used the sd-master-e410aeb-bin-win-avx2-x64 binary because using CPU-Z I saw my processor has AVX2 instruction. I tryed before the vulkan binary option, but it didn't work. I don't know why anymore... well, I just want to share that it took almost six hours to process the example command in README.md and it worked perfect, generating the same image. But my questions are:
Below, my command prompt screen. Thanks and great job with this project |
Beta Was this translation helpful? Give feedback.
Replies: 6 comments 1 reply
-
Vulkan won't work if you can't fit the whole model + latent image buffer in video memory, so it's not surprising it refused to work on your laptop. SD3 is a fairly big model compared to others like SD 1.5 and SDXL, the diffusion model alone is over 4 GB in f16, and with the text encoders it's over 16GB. One thing you could try to impove the generation time is to use a quantized version of the model. This will definitely reduce its memory footprint and maybe improve generaion time a bit (don't expect miracles). |
Beta Was this translation helpful? Give feedback.
-
There are a couple of things you can try:
|
Beta Was this translation helpful? Give feedback.
-
Thank you both @Green-Sky @stduhpf I'm very noob on LLM for generative images, so I looked for chatgpt to give me some clarifications on what are in your answers. So, I could try just change the model parameter to something like https://huggingface.co/second-state/stable-diffusion-3-medium-GGUF/blob/main/sd3-medium-Q8_0.gguf, right? Is there some quantized option to FLUX.dev? |
Beta Was this translation helpful? Give feedback.
-
Actually I think I found both of my answers on README.md. I can use for flux this link: https://huggingface.co/leejet/FLUX.1-dev-gguf?show_file_info=flux1-dev-q8_0.gguf and in the https://github.com/leejet/stable-diffusion.cpp/blob/master/docs/flux.md#convert-flux-weights there is an example using gguf file as model Thanks once more. |
Beta Was this translation helpful? Give feedback.
-
Use quantized lcm with TAESD in cudart or just buy new gear instead of torturing your potato. |
Beta Was this translation helpful? Give feedback.
-
Thanks @mchome 😅 I try only with the parameter which reduce the steps and the final image was very low resolution. Indeed it all points that I have to buy another machine. Anyone knows how is the suppor for those new Microsoft Copilot+ brand? New Copilot+ PCs from the biggest brands available... stands more in the end of the article... |
Beta Was this translation helpful? Give feedback.
Vulkan won't work if you can't fit the whole model + latent image buffer in video memory, so it's not surprising it refused to work on your laptop.
SD3 is a fairly big model compared to others like SD 1.5 and SDXL, the diffusion model alone is over 4 GB in f16, and with the text encoders it's over 16GB. One thing you could try to impove the generation time is to use a quantized version of the model. This will definitely reduce its memory footprint and maybe improve generaion time a bit (don't expect miracles).