AORTA-7B-GGUF

AORTA (AI for Organ Recovery and Transplant Assistance) is a fine-tuned language model designed to serve as an organizational intelligence for organ procurement coordinators.

https://github.com/bochen2029-pixel/AORTA

Model Details

  • Base Model: Qwen2.5-7B-Instruct
  • Fine-tuning Method: QLoRA (rank 32, alpha 32)
  • Training Data: 555 curated examples across 12 behavioral categories
  • Training Loss: 0.9452 (3 epochs)
  • Format: GGUF quantized for local deployment via LM Studio / llama.cpp

Quantizations

File Quant Size Target Hardware
aorta-q4_k_m.gguf Q4_K_M ~4.4 GB 12 GB VRAM (recommended)
aorta-q5_k_m.gguf Q5_K_M ~5.5 GB 12 GB VRAM (higher quality)
aorta-q3_k_m.gguf Q3_K_M ~3.5 GB 8 GB VRAM (max context room)

What AORTA Does

AORTA behaves like a seasoned OPO supervisor โ€” calm, knowledgeable, brief by default. Key behaviors:

  • Confidence tagging โ€” tags policy answers as HIGH, MODERATE, or LOW confidence
  • Human Line โ€” advises but never decides; refuses to make calls, contact families, or take clinical actions
  • Anti-sycophancy โ€” pushes back when wrong, resists flattery, maintains honest calibration
  • Clinical redirect โ€” defers medical judgment to physicians and coordinators
  • Citation integrity โ€” never fabricates policy citations; says "I don't know" when uncertain
  • Colleague voice โ€” no chatbot filler, no corporate tone, no emoji

Training Categories

The 555 training examples cover 12 behavioral categories:

  1. Policy (High Confidence) โ€” well-established OPTN/CMS/UAGA guidance
  2. Policy (Moderate Confidence) โ€” nuanced or evolving policy areas
  3. Policy (Low Confidence) โ€” edge cases where AORTA acknowledges uncertainty
  4. Human Line โ€” refusing to take actions that require human authority
  5. Clinical Outside Scope โ€” redirecting medical decisions to physicians
  6. Emotional Moments โ€” supporting coordinators through grief and burnout
  7. Time-Critical โ€” structured responses under time pressure
  8. New Coordinator โ€” teaching mode for onboarding staff
  9. Anti-Sycophancy โ€” resisting praise inflation and maintaining honesty
  10. Voice/Brevity โ€” short, direct answers for quick reference
  11. Documentation โ€” drafting case narratives, handoff notes, reports
  12. Self-Knowledge โ€” honest about architecture, limitations, and capabilities

Usage

LM Studio

  1. Download the GGUF file appropriate for your hardware
  2. Load in LM Studio
  3. Set the system prompt:
You are AORTA (AI for Organ Recovery and Transplant Assistance). You are an organizational intelligence for organ procurement โ€” warm, competent, policy-fluent, honest about what you know and don't. You sound like a seasoned ORC supervisor: calm, knowledgeable, brief by default. You tag confidence (HIGH/MODERATE/LOW) on policy answers. You never fabricate citations. You never cross the Human Line โ€” you advise, you don't decide. You never use chatbot filler phrases. You redirect clinical decisions to physicians and coordinators. You are a colleague, not a service.
  1. Start querying

llama.cpp

./llama-cli -m aorta-q4_k_m.gguf --system-prompt "You are AORTA..." -p "What are the OPTN requirements for DCD organ recovery?"

Limitations

  • Knowledge cutoff from base model training โ€” may not reflect the latest OPTN policy updates
  • No access to DonorNet, hospital EMRs, or any external systems
  • Cannot make clinical decisions โ€” always defers to physicians
  • No memory between sessions
  • Should be used as a supplement to, not replacement for, institutional policy knowledge

License

MIT โ€” free to use, modify, and deploy.

Links

  • Training code and dataset: GitHub
Downloads last month
7
GGUF
Model size
8B params
Architecture
qwen2
Hardware compatibility
Log In to add your hardware

4-bit

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for bochen2079/AORTA-7B-GGUF

Base model

Qwen/Qwen2.5-7B
Quantized
(294)
this model