🧠 Jan-v1-2509 GGUFs

Quantized version of: janhq/Jan-v1-2509


πŸ“¦ Available GGUFs

Format Description
F16 Full precision (16-bit), better quality, larger size βš–οΈ
Q8_K_XL Quantized (8-bit XL variant, based on the quantization table of the unsloth model Qwen3-4B-Thinking-2507), medium size, faster inference ⚑
Q4_K_XL Quantized (4-bit XL variant, based on the quantization table of the unsloth model Qwen3-4B-Thinking-2507), smaller size, faster inference ⚑

πŸš€ Usage

Example with llama.cpp:

./main -m ./gguf-file-name.gguf -p "Hello world!"
Downloads last month
23
GGUF
Model size
4B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

4-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for rodrigomt/Jan-v1-2509-GGUF

Finetuned
janhq/Jan-v1-2509
Quantized
(6)
this model