You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
LocalAI version:
2.23.0, podman installation, tried with latest-gpu-nvidia-cuda-12 and latest-aio-gpu-nvidia-cuda-12
` Environment, CPU architecture, OS, and Version:
Standard Debian 12 (96 GB memory) with two Nvidia Tesla P40 gpus (each 24 GB memory)
Describe the bug
LocalAI fails to load custom .gguf files. In this case its dolphin-2.5-mixtral-8x7b.Q5_K_M.gguf. Following error message appears with all backends
ERR [llama-cpp] Failed loading model, trying with fallback 'llama-cpp-fallback', error: failed to load model with internal loader: could not load model: rpc error: code = Canceled desc =
It loads and works without issue on version 2.22.1.
To Reproduce
Update LocalAI images to version 2.23.0 and load the "custom" model dolphin-2.5-mixtral-8x7b.Q5_K_M.gguf.
Expected behavior
Loading should work.
Logs
Multiple lines of following messages for all backends. ERR [llama-cpp] Failed loading model, trying with fallback 'llama-cpp-fallback', error: failed to load model with internal loader: could not load model: rpc error: code = Canceled desc =
The text was updated successfully, but these errors were encountered:
LocalAI version:
2.23.0, podman installation, tried with
latest-gpu-nvidia-cuda-12
andlatest-aio-gpu-nvidia-cuda-12
`
Environment, CPU architecture, OS, and Version:
Standard Debian 12 (96 GB memory) with two Nvidia Tesla P40 gpus (each 24 GB memory)
Describe the bug
LocalAI fails to load custom .gguf files. In this case its
dolphin-2.5-mixtral-8x7b.Q5_K_M.gguf
. Following error message appears with all backendsERR [llama-cpp] Failed loading model, trying with fallback 'llama-cpp-fallback', error: failed to load model with internal loader: could not load model: rpc error: code = Canceled desc =
It loads and works without issue on version 2.22.1.
To Reproduce
Update LocalAI images to version 2.23.0 and load the "custom" model
dolphin-2.5-mixtral-8x7b.Q5_K_M.gguf
.Expected behavior
Loading should work.
Logs
Multiple lines of following messages for all backends.
ERR [llama-cpp] Failed loading model, trying with fallback 'llama-cpp-fallback', error: failed to load model with internal loader: could not load model: rpc error: code = Canceled desc =
The text was updated successfully, but these errors were encountered: