Qwen2.5-7B-Instruct Q4_K_M GGUF

Quantized from Junn17/qwen using Unsloth.

Usage

llama-cli --model qwen_model.Q4_K_M.gguf -p "Hello!"
Downloads last month
-
GGUF
Model size
8B params
Architecture
qwen2
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Junn17/quantize_qwen

Base model

Qwen/Qwen2.5-7B
Quantized
(279)
this model