GGUF
conversational

Not working for me

#1
by ssp-namo - opened

https://github.com/ggml-org/llama.cpp
./llama-server -hf EssentialAI/rnj-1-instruct-GGUF --port 8000 --gpu-layers 99

I am getting this log output

llama_model_load: error loading model: error loading model architecture: unknown model architecture: 'rnj1'
llama_model_load_from_file_impl: failed to load model
common_init_from_params: failed to load model '/root/.cache/llama.cpp/EssentialAI_rnj-1-instruct-GGUF_rnj-1-8B-instruct-Q4_K_M.gguf', try reducing --n-gpu-layers if you're running out of VRAM
srv    load_model: failed to load model, '/root/.cache/llama.cpp/EssentialAI_rnj-1-instruct-GGUF_rnj-1-8B-instruct-Q4_K_M.gguf'
srv    operator(): operator(): cleaning up before exit...
main: exiting due to model loading error
Essential AI org

Hi, support for Rnj-1 is not yet in the official release of llama.cpp. Check the model card in this repo for how to build from the PR that we submitted.

Sign up or log in to comment