This model performs worse than the Mistral-Small-3.1-24B model with a 4-bit quantization.

#6
by zletpm - opened

I’m using a prompt to convert an MD file to plain text. However, this model performs worse than the Mistral-Small-3.1-24B model with a 4-bit quantization.

  1. The model follows fewer instructions. When prompted to return the processed plain text only, it provides “here is processed chunks:” or explanations like “I have done what….” instead.

  2. Repeating the generation, this 4-bit model gives a much higher probability of generating duplicated text.

Sign up or log in to comment