Localai model load fails w/GLM-4.7-Flash-IQ4_XS.gguf

#29
by SplashMatic - opened

ERROR Failed to load model modelID="GLM-4.7-Flash-IQ4_XS.gguf" error=failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF backend="llama-cpp"

i have used the two part f16 gguf successfully (1t/s) but it works...

Sign up or log in to comment