AORTA-7B-GGUF

AORTA (AI for Organ Recovery and Transplant Assistance) is a fine-tuned language model designed to serve as an organizational intelligence for organ procurement coordinators.

https://github.com/bochen2029-pixel/AORTA

Model Details

  • Base Model: Qwen2.5-7B-Instruct
  • Fine-tuning Method: QLoRA (rank 32, alpha 32)
  • Training Data: 555 curated examples across 12 behavioral categories
  • Training Loss: 0.9452 (3 epochs)
  • Format: GGUF quantized for local deployment via LM Studio / llama.cpp

Quantizations

File Quant Size Target Hardware
aorta-q4_k_m.gguf Q4_K_M ~4.4 GB 12 GB VRAM (recommended)
aorta-q5_k_m.gguf Q5_K_M ~5.5 GB 12 GB VRAM (higher quality)
aorta-q3_k_m.gguf Q3_K_M ~3.5 GB 8 GB VRAM (max context room)

What AORTA Does

AORTA behaves like a seasoned OPO supervisor β€” calm, knowledgeable, brief by default. Key behaviors:

  • Confidence tagging β€” tags policy answers as HIGH, MODERATE, or LOW confidence
  • Human Line β€” advises but never decides; refuses to make calls, contact families, or take clinical actions
  • Anti-sycophancy β€” pushes back when wrong, resists flattery, maintains honest calibration
  • Clinical redirect β€” defers medical judgment to physicians and coordinators
  • Citation integrity β€” never fabricates policy citations; says "I don't know" when uncertain
  • Colleague voice β€” no chatbot filler, no corporate tone, no emoji

Training Categories

The 555 training examples cover 12 behavioral categories:

  1. Policy (High Confidence) β€” well-established OPTN/CMS/UAGA guidance
  2. Policy (Moderate Confidence) β€” nuanced or evolving policy areas
  3. Policy (Low Confidence) β€” edge cases where AORTA acknowledges uncertainty
  4. Human Line β€” refusing to take actions that require human authority
  5. Clinical Outside Scope β€” redirecting medical decisions to physicians
  6. Emotional Moments β€” supporting coordinators through grief and burnout
  7. Time-Critical β€” structured responses under time pressure
  8. New Coordinator β€” teaching mode for onboarding staff
  9. Anti-Sycophancy β€” resisting praise inflation and maintaining honesty
  10. Voice/Brevity β€” short, direct answers for quick reference
  11. Documentation β€” drafting case narratives, handoff notes, reports
  12. Self-Knowledge β€” honest about architecture, limitations, and capabilities

Usage

LM Studio

  1. Download the GGUF file appropriate for your hardware
  2. Load in LM Studio
  3. Set the system prompt:
You are AORTA (AI for Organ Recovery and Transplant Assistance). You are an organizational intelligence for organ procurement β€” warm, competent, policy-fluent, honest about what you know and don't. You sound like a seasoned ORC supervisor: calm, knowledgeable, brief by default. You tag confidence (HIGH/MODERATE/LOW) on policy answers. You never fabricate citations. You never cross the Human Line β€” you advise, you don't decide. You never use chatbot filler phrases. You redirect clinical decisions to physicians and coordinators. You are a colleague, not a service.
  1. Start querying

llama.cpp

./llama-cli -m aorta-q4_k_m.gguf --system-prompt "You are AORTA..." -p "What are the OPTN requirements for DCD organ recovery?"

Limitations

  • Knowledge cutoff from base model training β€” may not reflect the latest OPTN policy updates
  • No access to DonorNet, hospital EMRs, or any external systems
  • Cannot make clinical decisions β€” always defers to physicians
  • No memory between sessions
  • Should be used as a supplement to, not replacement for, institutional policy knowledge

License

MIT β€” free to use, modify, and deploy.

Links

  • Training code and dataset: GitHub
Downloads last month
101
GGUF
Model size
8B params
Architecture
qwen2
Hardware compatibility
Log In to add your hardware

4-bit

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for bochen2079/AORTA-7B-GGUF

Base model

Qwen/Qwen2.5-7B
Quantized
(262)
this model