One-Token Rollout: Guiding Supervised Fine-Tuning of LLMs with Policy Gradient Paper • 2509.26313 • Published Sep 30 • 4
Parameter-Efficient Sparsity Crafting from Dense to Mixture-of-Experts for Instruction Tuning on General Tasks Paper • 2401.02731 • Published Jan 5, 2024 • 3
GroveMoE Collection GroveMoE is an open-source family of large language models developed by the AGI Center, Ant Research Institute. • 4 items • Updated Oct 13 • 7
Grove MoE: Towards Efficient and Superior MoE LLMs with Adjugate Experts Paper • 2508.07785 • Published Aug 11 • 28
Tulu 3 Datasets Collection All datasets released with Tulu 3 -- state of the art open post-training recipes. • 33 items • Updated about 19 hours ago • 95