Llama3.2-3B-Explained

A fine-tuned version of meta-llama/Llama-3.2-3B-Instruct trained on Explained 0.41k alpaca data using Auto-SFT — an automated hyperparameter search and supervised fine-tuning pipeline.

The base model was adapted to follow the style and content of the Explained 0.41k alpaca dataset. Expect improved performance on tasks similar to those represented in the training data.

Model Details

Property Value
Base model meta-llama/Llama-3.2-3B-Instruct
Training data data/Explained-0.41k-alpaca.json
Fine-tuning epochs 2
Fine-tuning date 2026-03-25
Fine-tuning method LoRA (merged to full 16-bit)

Training Hyperparameters

LoRA

Parameter Value
r 4
alpha 8
dropout 0.0
target_modules ['q_proj', 'v_proj', 'k_proj', 'o_proj']

Training

Parameter Value
learning_rate 1e-05
batch_size 1
gradient_accumulation_steps 2
warmup_ratio 0.0
max_seq_length 512

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model     = AutoModelForCausalLM.from_pretrained("theprint/Llama3.2-3B-Explained")
tokenizer = AutoTokenizer.from_pretrained("theprint/Llama3.2-3B-Explained")

Generated by Auto-SFT

Downloads last month
581
Safetensors
Model size
3B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for theprint/Llama3.2-3B-Explained

Adapter
(656)
this model
Adapters
3 models