230B vs 235B: Why no comparison against Qwen3-235B-A22B-Thinking-2507 ?
Why in your model card there's no direct comparison against a very similarly/same sized Qwen3-235B-A22B-Thinking-2507?
(I saw your reference to it being benchmarked by ArtificialAnalysis: artificialanalysis.ai/?models=minimax-m2%2Cglm-4-6-reasoning%2Cqwen3-235b-a22b-instruct-2507-reasoning)
PS; In the comments I saw people comparing MiniMax-M2 against GLM-4.6 (maybe because you also compare it in your model card), but AFAICSee, most say that GLM-4.6 performed better for their task; well, this is no surprise, as GLM-4.6 is a much larger, 357B, LLM.
MiniMax M2 is better then Qwen 235B A22B Thinking 2507(the new one), I'm pretty sure.
There are indications that show that the opposite can also be true.
https://livebench.ai/
https://www.youtube.com/watch?v=XHbuFRupSvk