-
Notifications
You must be signed in to change notification settings - Fork 100
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Crash on an endbr64
instruction.
#24
Comments
Hi! Please try to build and run llama.cpp and see if it works. If it crashes too with similar error, report the problem with If it does not crash, we would need to compare the code of |
I tried llama.cpp, and it worked without a crash. |
I took a look at As for the issue, I don't have any ideas how to fix this. |
I tried adding compile flags It doesn't make sense. I roughly read the code but didn't see anything close to that. The disassembly looks rather real, not like some random data. I'm dooooomed. |
@RnMss I've updated |
It still doest not work on my CPU. I'll try on Windows later. |
Got the same problem in docker nvcr.io/nvidia/pytorch:23.05-py3, tokenizers-0.13.3 |
try recompile the repo with disable the AVX instruction flag on cmakelist.txt @RnMss . this step works for me |
My build crashes inferencing with a model with "Illegal Instruction".
I debugged it and seems to crash on an
endbr64
instruction. I think my CPU doesn't support the instruction set.Is there a building option to turn off the instruction set?
Version: Master, commit e84c446d9533dabef2d8d60735d5924db63362ff
Command to reproduce
python rwkv/chat_with_bot.py ../models/xxxxxxx.bin
It crashed with "Illegal Instruction"
I debugged the program:
The text was updated successfully, but these errors were encountered: