This model performs worse than the Mistral-Small-3.1-24B model with a 4-bit quantization.
#6
by
zletpm
- opened
I’m using a prompt to convert an MD file to plain text. However, this model performs worse than the Mistral-Small-3.1-24B model with a 4-bit quantization.
The model follows fewer instructions. When prompted to return the processed plain text only, it provides “here is processed chunks:” or explanations like “I have done what….” instead.
Repeating the generation, this 4-bit model gives a much higher probability of generating duplicated text.