You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I followed the steps as mentioned, but when running Tinychat I repeatedly encountered an issue:
Initially, I tried to "Perform the AWQ search" but ran Out of Memory & the process was "Killed".
Then I tried to run the pre-saved results instead to run Tiny Chat [ .pt files provided in awq_cache] but eventually ran into the same issue ("Killed")
I have tried these solutions to resolve the issue, but none were helpful:
I added swap space of 16GB and then 64GB, but the process was still killed. I suspect that it may have timed out and was eventually terminated.
I then reduced the quantization group size from 128 to 64, but this did not resolve the issue either.
Sometimes these solutions work but eventually get killed after a few conversations.Could you please suggest the possible solution or optimization to resolve this memory issue? I've attached a screen shot for better understanding.
The text was updated successfully, but these errors were encountered:
Hi, it seems you did not activate flash attention here. The memory of Nano is very restricted. When sequence length gets larger , the memory consumption of attention grows quadratically (if flash_attention is not enabled). That's why the program runs into OOM after several rounds of conversation. We also do not suggest doing quantization on Jetson Orin Nano.
Thank you for the great work, I really appreciate it.
I followed the steps as mentioned, but when running Tinychat I repeatedly encountered an issue:
I have tried these solutions to resolve the issue, but none were helpful:
Sometimes these solutions work but eventually get killed after a few conversations.Could you please suggest the possible solution or optimization to resolve this memory issue? I've attached a screen shot for better understanding.
The text was updated successfully, but these errors were encountered: