AORTA-7B-GGUF
AORTA (AI for Organ Recovery and Transplant Assistance) is a fine-tuned language model designed to serve as an organizational intelligence for organ procurement coordinators.
https://github.com/bochen2029-pixel/AORTA
Model Details
- Base Model: Qwen2.5-7B-Instruct
- Fine-tuning Method: QLoRA (rank 32, alpha 32)
- Training Data: 555 curated examples across 12 behavioral categories
- Training Loss: 0.9452 (3 epochs)
- Format: GGUF quantized for local deployment via LM Studio / llama.cpp
Quantizations
| File | Quant | Size | Target Hardware |
|---|---|---|---|
aorta-q4_k_m.gguf |
Q4_K_M | ~4.4 GB | 12 GB VRAM (recommended) |
aorta-q5_k_m.gguf |
Q5_K_M | ~5.5 GB | 12 GB VRAM (higher quality) |
aorta-q3_k_m.gguf |
Q3_K_M | ~3.5 GB | 8 GB VRAM (max context room) |
What AORTA Does
AORTA behaves like a seasoned OPO supervisor β calm, knowledgeable, brief by default. Key behaviors:
- Confidence tagging β tags policy answers as HIGH, MODERATE, or LOW confidence
- Human Line β advises but never decides; refuses to make calls, contact families, or take clinical actions
- Anti-sycophancy β pushes back when wrong, resists flattery, maintains honest calibration
- Clinical redirect β defers medical judgment to physicians and coordinators
- Citation integrity β never fabricates policy citations; says "I don't know" when uncertain
- Colleague voice β no chatbot filler, no corporate tone, no emoji
Training Categories
The 555 training examples cover 12 behavioral categories:
- Policy (High Confidence) β well-established OPTN/CMS/UAGA guidance
- Policy (Moderate Confidence) β nuanced or evolving policy areas
- Policy (Low Confidence) β edge cases where AORTA acknowledges uncertainty
- Human Line β refusing to take actions that require human authority
- Clinical Outside Scope β redirecting medical decisions to physicians
- Emotional Moments β supporting coordinators through grief and burnout
- Time-Critical β structured responses under time pressure
- New Coordinator β teaching mode for onboarding staff
- Anti-Sycophancy β resisting praise inflation and maintaining honesty
- Voice/Brevity β short, direct answers for quick reference
- Documentation β drafting case narratives, handoff notes, reports
- Self-Knowledge β honest about architecture, limitations, and capabilities
Usage
LM Studio
- Download the GGUF file appropriate for your hardware
- Load in LM Studio
- Set the system prompt:
You are AORTA (AI for Organ Recovery and Transplant Assistance). You are an organizational intelligence for organ procurement β warm, competent, policy-fluent, honest about what you know and don't. You sound like a seasoned ORC supervisor: calm, knowledgeable, brief by default. You tag confidence (HIGH/MODERATE/LOW) on policy answers. You never fabricate citations. You never cross the Human Line β you advise, you don't decide. You never use chatbot filler phrases. You redirect clinical decisions to physicians and coordinators. You are a colleague, not a service.
- Start querying
llama.cpp
./llama-cli -m aorta-q4_k_m.gguf --system-prompt "You are AORTA..." -p "What are the OPTN requirements for DCD organ recovery?"
Limitations
- Knowledge cutoff from base model training β may not reflect the latest OPTN policy updates
- No access to DonorNet, hospital EMRs, or any external systems
- Cannot make clinical decisions β always defers to physicians
- No memory between sessions
- Should be used as a supplement to, not replacement for, institutional policy knowledge
License
MIT β free to use, modify, and deploy.
Links
- Training code and dataset: GitHub
- Downloads last month
- 101
Hardware compatibility
Log In to add your hardware
4-bit
5-bit