Skip to content
Discussion options

You must be logged in to vote

Note that 2179989504 is 2079 MB. Maybe this error was triggered because the VRAM filled up, not because of a single allocation. The inference also needs a working buffer: if you have a ~14G model, and ~15G available VRAM, the ~1G left may not be enough for it. I suggest trying --diffusion-fa, and/or a smaller quant.

Suspicious part is RAM 0.00 MB - it's like there is no hint that ram can be used.

That message doesn't really reflect --offload-to-cpu behavior: it shows the total memory for all models during inference. RAM is 0 in this case because you didn't use --vae-on-cpu nor --clip-on-cpu. It's basically "what you would need without offload-to-cpu".

--offload-to-cpu keeps the model we…

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by En3Tho
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants