Not working for me
#1
by
ssp-namo
- opened
https://github.com/ggml-org/llama.cpp
./llama-server -hf EssentialAI/rnj-1-instruct-GGUF --port 8000 --gpu-layers 99
I am getting this log output
llama_model_load: error loading model: error loading model architecture: unknown model architecture: 'rnj1'
llama_model_load_from_file_impl: failed to load model
common_init_from_params: failed to load model '/root/.cache/llama.cpp/EssentialAI_rnj-1-instruct-GGUF_rnj-1-8B-instruct-Q4_K_M.gguf', try reducing --n-gpu-layers if you're running out of VRAM
srv load_model: failed to load model, '/root/.cache/llama.cpp/EssentialAI_rnj-1-instruct-GGUF_rnj-1-8B-instruct-Q4_K_M.gguf'
srv operator(): operator(): cleaning up before exit...
main: exiting due to model loading error
Hi, support for Rnj-1 is not yet in the official release of llama.cpp. Check the model card in this repo for how to build from the PR that we submitted.