-
-
Notifications
You must be signed in to change notification settings - Fork 36
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
not enough space in the context's memory pool (on Apple M1 Max, 32GB RAM, clip-vit-b-32) #33
Comments
Barely missing the threshold on ggml_new_tensor_impl: not enough space in the context's memory pool (needed 17471536, available 16777216)
Assertion failed: (false), function ggml_new_tensor_impl, file ggml.c, line 4449.
zsh: abort ./main --model --text "test" --image -v 1 |
It requests ~12 mb instead of 8mb that I set as a fixed value here. Lines 24 to 30 in e2eee8e
You can slightly increase them --8 is for patch32 and 16 is for patch16, so adjust them to a value that is working for you. Interestingly, It works for me with these values on Windows and Linux but haven't tried with Macbook yet. Additionally, quantized models may require slightly more memory. I'll try to replicate it tomorrow.
What kind of devices are you targeting? I'm quite interested in new use cases and low-end devices, so we can work on it anyway |
That worked! Thank you so much for the quick reply. I want to run on Intel-era Macbook Pros and Airs, like a 13" Macbook Air 2019. Not very low-end in the grand scheme of things haha |
As an aside, have you tried converting these models to CoreML like they do in whisper.cpp? |
That's great! I'll try to find the root cause of this difference and patch it later on.
Hahha yes. They should do a fairly good job.
Not yet, but good point. I'd like to support additional deployment types as we find different use cases for clip.cpp. |
Really appreciate the quick replies here. Have you also considered building out a version of this for BLIP or other more recent CLIP variants? Currently exploring the steps involved. Large-scale image retrieval has worked far better on BLIP and BLIP2 but of course they take way more time and memory |
Hi there,
Thank you so much for making this library. I'm unfortunately running into the following error
I'm running on a Mac Studio with M1 Max and 32 GB of RAM. I tried every available model binary on huggingface and still got the same memory pool error. Is this due to a memory allocation bug? I see in #17 that this got solved for some cases and I'm wondering if there are lingering issues here
The text was updated successfully, but these errors were encountered: