Config file not being read #1139
Unanswered
gurbindersingh
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I've set up LocalAI on a Ubuntu 22.04 machine with a Nvidia GPU following the easy setup guide. But it looks like there is no GPU offloading.
I can see the following message in the logs
offloading 0/43 layers to the GPU
andVRAM used: 0MB
.It took a while to figure out that you need the
gpu_layers
parameter in the config file since that's not mentioned in the guide. But I've now created the config file in the same directory as the model (official Llama 2 model converted using Llama.cpp). This is what it looks like:But it doesn't seem like these configs are being read, even after restarting (and even rebuilding) the container. I am not sure if this is a bug or if I'm doing something wrong since the docs don't mention which of the properties in the configs are required and which are not.
Edit: I can run the model just fine with GPU offloading using Llama.cpp at 70 tokens/s. With LocalAI it's only running at 5-7 tokens/s.
Beta Was this translation helpful? Give feedback.
All reactions