Post
453
MiniCPM4🔥 efficient LLMs built for end-side devices, by OpenBMB
openbmb/minicpm4-6841ab29d180257e940baa9b
✨ Apache 2.0
✨ 5–7× Faster Inference (Jetson Orin & RTX 4090)
✨ 8B trained on 8T clean, non-synthetic tokens
✨ 32K Native Context -> 128K+ with InfLLM v2 + LongRoPE
✨ Runs on 🤗Transformers , http://CPM.cu, vLLM, and SGLang
openbmb/minicpm4-6841ab29d180257e940baa9b
✨ Apache 2.0
✨ 5–7× Faster Inference (Jetson Orin & RTX 4090)
✨ 8B trained on 8T clean, non-synthetic tokens
✨ 32K Native Context -> 128K+ with InfLLM v2 + LongRoPE
✨ Runs on 🤗Transformers , http://CPM.cu, vLLM, and SGLang