You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi! With the new version of Forge, and FLUX, this extension could be really practical for the millions of low VRAM laptops that can now run FLUX. The only problem is that it doesn't unload the LLM from VRAM when using Ollama, so the generation is way too slow.
Also, could it be set to store a different system prompt and ollama settings? I found that giving it an example in the system prompt works well.
Basically, it seems you are a few lines of code away from the best 'magic prompt' software in the world, surpassing the ones on Dalle-3 and Ideogram, which are censored. Thank you!
The text was updated successfully, but these errors were encountered:
Hi! With the new version of Forge, and FLUX, this extension could be really practical for the millions of low VRAM laptops that can now run FLUX. The only problem is that it doesn't unload the LLM from VRAM when using Ollama, so the generation is way too slow.
According to ollama/ollama#1600 , that can be accomplished with
'''
curl http://localhost:11434/api/generate -d '{"model": "llama2", "keep_alive": 0}'
'''
Can that be put in your code?
Also, could it be set to store a different system prompt and ollama settings? I found that giving it an example in the system prompt works well.
Basically, it seems you are a few lines of code away from the best 'magic prompt' software in the world, surpassing the ones on Dalle-3 and Ideogram, which are censored. Thank you!
The text was updated successfully, but these errors were encountered: