Qwen3-ASR-1.7B-GGUF

This model is converted from Qwen/Qwen3-ASR-1.7B to GGUF using convert_hf_to_gguf.py

To use it:

llama-server -hf ggml-org/Qwen3-ASR-1.7B-GGUF
Downloads last month
2,310
GGUF
Model size
2B params
Architecture
qwen3vl
Hardware compatibility
Log In to add your hardware

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ggml-org/Qwen3-ASR-1.7B-GGUF

Quantized
(17)
this model