Inference Providers
Active filters: glm4_moe
cyankiwi/GLM-4.5-Air-AWQ-4bit
Text Generation
• 19B • Updated • 4.94k
• 28
mlx-community/GLM-4.5-Air-3bit-DWQ
Text Generation
• 107B • Updated • 77
• 4
QuantTrio/GLM-4.5-Air-AWQ-FP16Mix
Text Generation
• 24B • Updated • 1.49k
• 14
QuantTrio/GLM-4.5-Air-GPTQ-Int4-Int8Mix
Text Generation
• 20B • Updated • 310
• 10
QuantTrio/GLM-4.5-GPTQ-Int4-Int8Mix
Text Generation
• 55B • Updated • 86
• 5
OnFinanceAI/GLM-4.5-Air-FP4
107B • Updated • 4
• 1
Text Generation
• 53B • Updated • 30
• 9
mlx-community/GLM-4.5-Air-2bit-DWQ
Text Generation
• 107B • Updated • 26
• 2
Conexis/GLM-4.5-Air-Channel-INT8
Text Generation
• 111B • Updated • 4
• 1
Conexis/GLM-4.5-Channel-INT8
Text Generation
• 358B • Updated • 4
• 1
Doctor-Shotgun/GLM-4.5-Air-exl3_5.0bpw-h6
Text Generation
• 34B • Updated • 5
• 2
Text Generation
• 110B • Updated • 87
• 8
Text Generation
• 358B • Updated • 18
• 7
LatentWanderer/zai-org_GLM-4.5-Air-4.5bpw-h8-exl3
Text Generation
• 31B • Updated • 4
cyankiwi/GLM-4.5-Air-AWQ-8bit
Text Generation
• 32B • Updated • 126
• 2
bullerwins/GLM-4.5-exl3-3.0bpw
Text Generation
• 67B • Updated • 4
bullerwins/GLM-4.5-exl3-4.0bpw
Text Generation
• 89B • Updated • 3
Minthy/GLM-4.5-Air-exl3-6bpw
Text Generation
• 41B • Updated • 5
• 1
rocca/GLM-4.5-compressed-tensors-w4a16-g128-v1
48B • Updated • 1
Text Generation
• 358B • Updated • 4
Minthy/GLM-4.5-Air-exl3-5.5bpw
Text Generation
• 37B • Updated • 5
bullerwins/GLM-4.5-exl3-5.0bpw
Text Generation
• 111B • Updated • 7
monirmamoun/GLM-4.5-MLX-3bit
Text Generation
• 353B • Updated • 8
Doctor-Shotgun/GLM-4.5-Air-exl3_3.08bpw-h6
21B • Updated • 2
Doctor-Shotgun/GLM-4.5-Air-exl3_3.14bpw-h6
22B • Updated • 3
• 3
mlx-community/GLM-4.5-Air-3bit-DWQ-v2
Text Generation
• 107B • Updated • 31
• 4
monirmamoun/GLM-4.5-MLX-2bit
Text Generation
• 353B • Updated • 5
FlagRelease/GLM-4.5-FlagOS
bullerwins/GLM-4.5-exl3-3.5bpw
Text Generation
• 78B • Updated • 6
bullerwins/GLM-4.5-exl3-3.2bpw_optim
Text Generation
• 71B • Updated • 2