Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llama-cpp of LocaAI 2.23.0 doesn't load custom .gguf files #4199

Open
brknkfr opened this issue Nov 19, 2024 · 1 comment
Open

llama-cpp of LocaAI 2.23.0 doesn't load custom .gguf files #4199

brknkfr opened this issue Nov 19, 2024 · 1 comment
Labels
bug Something isn't working unconfirmed

Comments

@brknkfr
Copy link

brknkfr commented Nov 19, 2024

LocalAI version:
2.23.0, podman installation, tried with latest-gpu-nvidia-cuda-12 and latest-aio-gpu-nvidia-cuda-12
`
Environment, CPU architecture, OS, and Version:
Standard Debian 12 (96 GB memory) with two Nvidia Tesla P40 gpus (each 24 GB memory)

Describe the bug
LocalAI fails to load custom .gguf files. In this case its dolphin-2.5-mixtral-8x7b.Q5_K_M.gguf. Following error message appears with all backends

ERR [llama-cpp] Failed loading model, trying with fallback 'llama-cpp-fallback', error: failed to load model with internal loader: could not load model: rpc error: code = Canceled desc =

It loads and works without issue on version 2.22.1.

To Reproduce
Update LocalAI images to version 2.23.0 and load the "custom" model dolphin-2.5-mixtral-8x7b.Q5_K_M.gguf.

Expected behavior
Loading should work.

Logs
Multiple lines of following messages for all backends.
ERR [llama-cpp] Failed loading model, trying with fallback 'llama-cpp-fallback', error: failed to load model with internal loader: could not load model: rpc error: code = Canceled desc =

@brknkfr brknkfr added bug Something isn't working unconfirmed labels Nov 19, 2024
@brknkfr
Copy link
Author

brknkfr commented Nov 19, 2024

Maybe related to #4170

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working unconfirmed
Projects
None yet
Development

No branches or pull requests

1 participant