We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hello! I'm trying to run a model on my M4 mac mini but I always get this when requesting an answer:
error: Insufficient Memory (00000008:kIOGPUCommandBufferCallbackErrorOutOfMemory) llama_graph_compute: ggml_backend_sched_graph_compute_async failed with error -1 llama_decode: failed to decode, ret = -3
Failed to predict at 43! Check your context buffer sizes!
The text was updated successfully, but these errors were encountered:
Error says " Insufficient Memory". What gguf are you trying to load, and how much memory does your M4 mac mini have?
Sorry, something went wrong.
No branches or pull requests
Hello! I'm trying to run a model on my M4 mac mini but I always get this when requesting an answer:
error: Insufficient Memory (00000008:kIOGPUCommandBufferCallbackErrorOutOfMemory)
llama_graph_compute: ggml_backend_sched_graph_compute_async failed with error -1
llama_decode: failed to decode, ret = -3
Failed to predict at 43! Check your context buffer sizes!
The text was updated successfully, but these errors were encountered: