Papers
arxiv:2511.05650

Optimizing Diversity and Quality through Base-Aligned Model Collaboration

Published on Nov 7
· Submitted by Chenghao Yang on Nov 12
Authors:
,
,
,
,
,

Abstract

BACo, a token-level collaboration framework, enhances diversity and quality in large language model outputs by dynamically routing between a base model and its aligned counterpart.

AI-generated summary

Alignment has greatly improved large language models (LLMs)' output quality at the cost of diversity, yielding highly similar outputs across generations. We propose Base-Aligned Model Collaboration (BACo), an inference-time token-level model collaboration framework that dynamically combines a base LLM with its aligned counterpart to optimize diversity and quality. Inspired by prior work (Fei et al., 2025), BACo employs routing strategies that determine, at each token, from which model to decode based on next-token prediction uncertainty and predicted contents' semantic role. Prior diversity-promoting methods, such as retraining, prompt engineering, and multi-sampling methods, improve diversity but often degrade quality or require costly decoding or post-training. In contrast, BACo achieves both high diversity and quality post hoc within a single pass, while offering strong controllability. We explore a family of routing strategies, across three open-ended generation tasks and 13 metrics covering diversity and quality, BACo consistently surpasses state-of-the-art inference-time baselines. With our best router, BACo achieves a 21.3% joint improvement in diversity and quality. Human evaluations also mirror these improvements. The results suggest that collaboration between base and aligned models can optimize and control diversity and quality.

Community

Tired of aligned LLMs losing their creativity? 🤖
Alignment improves LLM quality but badly hurts output diversity. This "diversity-quality trade-off" forces a choice: do you want creative answers or high-quality ones?
What if you could have both?
Excited to share our new paper:
BACO (Base-Aligned Model Collaboration)!
BACO is a new inference-time framework that gets the best of both worlds. It dynamically "collaborates" between a base LLM (for high diversity) and its aligned counterpart (for high quality) at the token level.
🚀 The result: A 21.3% joint improvement in diversity & quality—all in a single pass with no costly retraining.

Code
Data
Awesome LLM Diversity Reading List

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2511.05650 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2511.05650 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2511.05650 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.