Model fails to run in llama.cpp unknown model architecture: 'vaetki'
Is there a specific llama build or PR required to run this model?
llama-cli -m VAETKI-GGUF\VAETKI-Q4_K_M.gguf -p "Your prompt here"
ggml_cuda_init: found 1 CUDA devices:
Device 0: NVIDIA GeForce RTX 5090, compute capability 12.0, VMM: yes
load_backend: loaded CUDA backend from S:\llm\llama\ggml-cuda.dll
load_backend: loaded RPC backend from S:\llm\llama\ggml-rpc.dll
load_backend: loaded CPU backend from S:\llm\llama\ggml-cpu-alderlake.dll
Loading model... -llama_model_load: error loading model: error loading model architecture: unknown model architecture: 'vaetki'
llama_model_load_from_file_impl: failed to load model
llama_params_fit: encountered an error while trying to fit params to free device memory: failed to load model \llama_model_load: error loading model: error loading model architecture: unknown model architecture: 'vaetki' llama_model_load_from_file_impl: failed to load model
common_init_from_params: failed to load model 'VAETKI-GGUF\VAETKI-Q4_K_M.gguf'
srv load_model: failed to load model, 'VAETKI-GGUF\VAETKI-Q4_K_M.gguf'
Failed to load the model
VAETKI support requires a modified version of llama.cpp. Until the PR is merged, please use: https://github.com/dororodoroddo/llama.cpp/tree/add-vaetki-support
I am currently addressing the change requests for this PR (as of Jan 10, 11:41 UTC). There might be some adjustments to the GGUF format, so please bear with me. I will upload the updated files and notify you as soon as they are ready.