-
Notifications
You must be signed in to change notification settings - Fork 212
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
llama 3.1 8b not working on SRP #2887
Comments
@dhandhalyabhavik The fix has be pushed to the main branch today. There was an issue with the tokenizer which got updated in the commit d44aeb3 |
Its fixed, Thanks @dtrawins. Closing the issue now. |
Hi @dtrawins Re-opening the issue because I am not able to make it working on NUC device. I tried 2 things, 1 first tried using already released 2024.5 ovms image, that resulted in same above error,
2 Tried building image from scratch and mid way resulted into this error,
My NUC details:
Please help me answer the question. Also if possible answer the previous unanswered question, |
Hi @dtrawins, can you please help? |
@dhandhalyabhavik If you are using the public image 2024.5, you should also export the models using the export tool from that release branch https://github.com/openvinotoolkit/model_server/tree/releases/2024/5/demos/common/export_models |
Describe the bug
while running model from ovms server docker image, its not running properly. Logs suggested there was problem with model conversion. So here I am providing all the logs for debug.
To Reproduce
Steps to reproduce the behavior:
Model Conversion
It has generated these files
Logs from docker image,
Configuration
cloned main repo today (29th Nov), build docker image (CPU) from scratch. And I faced above error.
Please help me find the issue.
Additional Recommendation/Thoughts
I am just wondering, why we are not pushing this converted models to Huggingface and pulling from there? Because this is not the first time I have faced issue with conversion of model to quantized version.
The text was updated successfully, but these errors were encountered: