Turkish LLM Family
Collection
Open-source Turkish LLM family (1.5B-32B). Models, GGUF quantizations, datasets, and demos. • 8 items • Updated
A Turkish-enhanced 14B model fine-tuned from Qwen2.5-14B-Instruct with QLoRA on 242K Turkish instruction examples.
Part of the Turkish LLM Family.
| Benchmark | Base (Qwen2.5-14B) | Ours | Delta |
|---|---|---|---|
| MMLU-TR | 0.5947 | 0.5977 | +0.30 |
ollama run hf.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF:Q4_K_M
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("ogulcanaydogan/Turkish-LLM-14B-Instruct", torch_dtype="auto", device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("ogulcanaydogan/Turkish-LLM-14B-Instruct")
messages = [
{"role": "system", "content": "Sen yardimci bir Turkce asistansin."},
{"role": "user", "content": "Yapay zekanin egitim sektorundeki etkilerini acikla."}
]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer([text], return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
| Parameter | Value |
|---|---|
| Base Model | Qwen/Qwen2.5-14B-Instruct |
| Method | QLoRA (4-bit NF4) |
| LoRA rank / alpha | 32 / 64 |
| Learning rate | 1e-5 |
| Dataset | 242K Turkish instruction examples |
| Model | Size | MMLU-TR | GGUF |
|---|---|---|---|
| Turkish-LLM-7B | 7B | - | Download |
| Turkish-LLM-14B | 14B | 0.5977 | Download |
| Turkish-LLM-32B | 32B | 0.6564 | Download |
@misc{aydogan2026turkishllm,
title={Turkish LLM Family: Open-Source Turkish Language Models},
author={Ogulcan Aydogan},
year={2026},
url={https://huggingface.co/collections/ogulcanaydogan/turkish-llm-family-69b303b4ef1c36caffca4e94}
}