Llama-2-7B-Chat Q4_K_M GGUF
Quantized from Junn17/llama using Unsloth.
Usage
llama-cli --model llama-2-7b-chat.Q4_K_M.gguf -p "Hello!"
- Downloads last month
- -
Hardware compatibility
Log In to add your hardware
4-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for Junn17/quantize_llama2
Base model
meta-llama/Llama-2-7b-chat-hf