Localai model load fails w/GLM-4.7-Flash-IQ4_XS.gguf
#29
by SplashMatic - opened
ERROR Failed to load model modelID="GLM-4.7-Flash-IQ4_XS.gguf" error=failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF backend="llama-cpp"
i have used the two part f16 gguf successfully (1t/s) but it works...