-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bug: llama 3.1 and variants fail with error "wrong number of tensors; expected 292, got 291" #516
Comments
not just windows -->all Linux flavors: |
See also SciSharp/LLamaSharp#874 I'm having the same issue with |
Same with |
Similar results with Llama-3.1-8b-Pruned-7-Layers.Q8_0.gguf albeit expected 220 got 219 |
This issue is probably fixed as of commit e9ee3f9 |
Could someone please provide step-by-step instructions to resolve this issue? I'm encountering the same problem and would appreciate guidance on how to fix it. |
Wait for the next binary release if you don't want to compile it yourself |
I tried building it locally from the current master but it isn't working
|
This issue has been fixed as of release v0.8.13. |
@camAtGitHub want to close the issue now that its fixed? |
Contact Details
github
What happened?
I came here to report the issue / bug / my incompetence around the error of:
llama_model_load: error loading model: done_getting_tensors: wrong number of tensors; expected 292, got 291
The logs are below but in particular trying to load external weights (on Windows) for
Meta-Llama-3.1-8B-Instruct-Q4_K_M.gguf
and its variant:DarkIdol-Llama-3.1-8B-Instruct-1.2-Uncensored.Q4_K_M.gguf
result in the above error about 'wrong number of tensors'.Some digging I found that LM studio v0.2.29 is required so I went looking for changes between LM studio 0.2.28 and .29 and at a pure guess it seems that llama 3.1 rope scaling was introduced / patched / fix, which maybe the reason for the error as llamafile doesn't have this yet? (again a guess on my part).
Anyway to reproduce the error it should be easy enough just try:
llamafile-0.8.12.exe -m Meta-Llama-3.1-8B-Instruct-Q4_K_M.gguf
(sha1sum ee8c490b5390f3d85e59b2b2c61d83157ce5df73) and see if you get the same error.Additional info
llamafile-0.8.12.exe -m mistral-7b-instruct-v0.1.Q4_K_M.gguf
works fine.llamafile
but does when usingllama-cli
TL;DR
Meta-Llama-3.1-8B-Instruct and variants arent running with
llamafile
failing with errorllama_model_load: error loading model: done_getting_tensors: wrong number of tensors; expected 292, got 291
.I believe the issue to be related to changes in the gguf file type/model that have been introduced.
Version
llamafile v0.8.12
What operating system are you seeing the problem on?
Windows
Relevant log output
The text was updated successfully, but these errors were encountered: