https://huggingface.co/nightmedia/Qwen3-14B-Scientist-BF16

#1650
by nightmedia - opened

Once again hopeful that it will work this time :)

https://huggingface.co/nightmedia/Qwen3-14B-Scientist-BF16

I added a few more models to make it fun

  • internlm/JanusCoder-14B
  • Azure99/Blossom-V6.3-14B
  • TeichAI/Qwen3-14B-Polaris-Alpha-Distill
  • TeichAI/Qwen3-14B-Gemini-3-Pro-Preview-High-Reasoning-Distill
  • TeichAI/Qwen3-14B-Claude-4.5-Opus-High-Reasoning-Distill
  • MegaScience/Qwen3-14B-MegaScience
  • Jasaxion/MathSmith-HC-Qwen3-14B-ShortCoT

Accept me, I will queue a bit later as I need to go somewhere right now, and queue is busy lol. If any more models let me know so I can queue when I come back

Well, since you are willing, here is a MoE that would be nice in GGUF

https://huggingface.co/nightmedia/Qwen3-30B-A3B-Architect7

GAIR/SR-Scientist-30B
NousResearch/nomos-1
YOYO-AI/Qwen3-30B-A3B-YOYO-V2
YOYO-AI/Qwen3-30B-A3B-YOYO-V4
miromind-ai/MiroThinker-v1.0-30B

I get these metrics in MLX

mxfp4    0.551,0.692,0.876,0.749,0.422,0.794,0.691
qx64-hi  0.561,0.725,0.879,0.753,0.468,0.794,0.686
qx86-hi  0.563,0.737,0.878,0.758,0.448,0.803,0.698

qx86-hi  PPL: 4.392 ± 0.026

It might probably taper off by q3 to act like a normal MoE

nightmedia/Qwen3-14B-Scientist-BF16: no architectures entry (malformed JSON string, neither tag, array, object, number, string or atom, at character offset 0 (before "(end of string)")

architect7 was added a week ago apparently: https://hf.tst.eu/model#Qwen3-30B-A3B-Architect7-i1-GGUF

not sure why json thing happens, it seems like you already have it, perhaps something is wrong with formatting or forgot a comma somewhere?

That’s okay, I will dig into why that happens, probably you have a strict set of rules and I am missing something… I’ll fix it.

nightmedia changed discussion status to closed

I mean I can try to force it, but because of the queue it will take a while to get an answer. Want to fix it or should I try to force it ?

No, give me a bit to double check. It might be my chat templates that I included—copy-pasted it, and might contain an “atom of garbage” :)
If I have the jinja template, there’s no need to put it in the config

I was also tired last night, I meant the other Architect
https://huggingface.co/nightmedia/Qwen3-30B-A3B-Architect5

This is a version without nomos, different performance profile

The MoEs had no issues, mergekit has problems with dense models… and probably me doing way too many things at once

Ok, I removed the chat templates from the tokenizer_config.json, it should be fine now, let’s try again 🙏😀

https://huggingface.co/nightmedia/Qwen3-30B-A3B-Architect5

HTTP 404, please make it public =)

nightmedia/Qwen3-14B-Scientist-BF16: no architectures entry (malformed JSON string, neither tag, array, object, number, string or atom, at character offset 0 (before "(end of string)")

that's so weird, let's ask @nicoboss

I found the dtype was set to float32, probably got carried over from a merge
Changed it to
"dtype": "bfloat16",

@RichardErkhov This is probably a side effect from queuing a gated model. The JSON syntax validator doesn't use the HuggingFace token you provide and so tries to validate the 404 error as JSON and complains. If you provide a Huggingface token you know has access to the model it is save to ignore it and simply force add it. Should the JSON really be invalid it will simply fail after downloading and you can review the issue using llmc audit. If the JSON is invalid the HuggingFace model card will usually also put a warning on top as it does for all the latest Nemotron models where NVidia puts a corrupted JSON on purpose despite knowing everyone hates them for doing so. Please also make sure to specify nico1 as worker for gated models as otherwise importance matrix computation will fail if it gets queued to rich1 or marco as thouse expect nico1 to redownload the model for importance matrix computation for which the HuggingFace token would be missing on nico1.

Ah yes, I forgot the token! How could I ! Ok 5 minutes I queue with token

Added both models, let's see how it goes

Sign up or log in to comment