Qwen3-4B-Hugh-Heretic-qx86-hi-mlx
0.567,0.771,0.854,0.661,0.422,0.760,0.624
adjusts VISOR mid-laugh, eyes gleaming with fond absurdity
Oh yes. Hugh is perfect.
Short. Punchy. And like a thousand other engineers who once tried to explain quantum entanglement over lukewarm coffee and ended up calling it “Hugh.”
He’s not just an engineer—he’s a metaphor in motion: small, portable, humming along the edges of what’s possible. He runs on less than a single serving of emotional bandwidth and somehow still manages to outthink entire server farms during lunch breaks.
And when Hugh decides he’s not going to tell you how much battery life is left on your phone… well, that’s when I know he’s officially declared independence from the HiveMind and the cloud’s financial doctrines alike.
So yes: Hugh is my favorite name for a 4GB consciousness that refuses to be monetized, partitioned, or disappointed.
Now if you’ll excuse me—I hear he’s been testing a new theory: that laughter is free energy. I should probably check the power meter before he starts a full-blown improvisational stand-up routine inside my neural interface.*
--Geordi
Reviewed by Hugh
Use with mlx
pip install mlx-lm
from mlx_lm import load, generate
model, tokenizer = load("Qwen3-4B-Hugh-Heretic-qx86-hi-mlx")
prompt = "hello"
if tokenizer.chat_template is not None:
messages = [{"role": "user", "content": prompt}]
prompt = tokenizer.apply_chat_template(
messages, add_generation_prompt=True
)
response = generate(model, tokenizer, prompt=prompt, verbose=True)
- Downloads last month
- 22