This is zai-org/GLM-4.7-Flash quantized with llm-compressor to FP8. The model is compatible with vLLM (tested: v0.14.0). Tested with an L4 (Google Colab).
- Developed by: The Kaitchup
- License: lfm1.0
- Downloads last month
- 76
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support
Model tree for kaitchup/GLM-4.7-Flash-FP8-Dynamic
Base model
zai-org/GLM-4.7-Flash