Austral 24B Winton

Overview
Austral 24B - Winton
More than 1.5-metres tall, about six-metres long and up to 1000-kilograms heavy, Australovenator Wintonensis was a fast and agile hunter. The largest known Australian theropod.
This is a finetune of Harbinger 24B to be a generalist Roleplay/Adventure model. I've removed some of the "slops" that i noticed in an otherwise great model aswell as improving the general writing of the model, This was a multi-stage finetune, all previous checkpoints are released aswell.
Support my finetunes / Me on Kofi: https://Ko-fi.com/deltavector | Thank you to Auri for helping/Testing ♥
Quants
Chat Format
This model utilizes ChatML.
<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
Training
As the the Austral/Francois tradition, I built off another great finetune Harbinger-24B, I did 4 epochs ontop with roughly the same datamix as Francois-Huali/Austral 70B as a R128 Lora, then KTO alignment with a mix of Instruct/Small writing datasets and then finally another 4 epoch SFT with Rep_remover (Thanks Pocket!)
Config(Post-KTO SFT)
https://wandb.ai/new-eden/austral/artifacts/axolotl-config/config-0tzehrhe/v0/files/axolotl_config_m8018fm4.yml
This model was trained over 4 epochs using 8 x A100s for the base SFT, Then i used KTO to clean up some coherency issues for 1 epoch, then finally training for another 4 epochs on Rep_Remover to delete slops. Total was roughly 80 hours total.
Credits
TYSM to my friends: Auri, Lucy, Trappu, Alicat, Kubernetes Bad, Intervitens, NyxKrage & Kalomaze
- Downloads last month
- 180
Model tree for Delta-Vector/Austral-24B-Winton
Base model
mistralai/Mistral-Small-3.1-24B-Base-2503