Please Gguf's

#2
by tech77 - opened

Please for Gguf's to work at Lmstudio.

To save space, you can quantify it yourselves.

I've got too poor hardware to make gguf's.

GGUF quantization does not require GPU memory.

Pleeeeaaaase... :))
We need gguf.

If it's okay I make request of making Gguf's of this model from mradermacher - https://huggingface.co/mradermacher/model_requests/discussions/1793

convert_hf_to_gguf fails with Can not map tensor 'model.vit_large_projector.weight after

INFO:hf-to-gguf:blk.35.attn_q_norm.weight, torch.bfloat16 --> F32, shape = {128}
INFO:hf-to-gguf:output_norm.weight,        torch.bfloat16 --> F32, shape = {4096}

convert_hf_to_gguf fails with Can not map tensor 'model.vit_large_projector.weight after

INFO:hf-to-gguf:blk.35.attn_q_norm.weight, torch.bfloat16 --> F32, shape = {128}
INFO:hf-to-gguf:output_norm.weight,        torch.bfloat16 --> F32, shape = {4096}

Hi, come to my repo for GGUF, you could build llama-server or llama-mtmd-cli yourself from my branch: https://github.com/Kraven1109/llama.cpp
it's quite fast on my laptop

Please for Gguf's to work at Lmstudio.

hi

Sign up or log in to comment