--- base_model: - Qwen/Qwen3-4B-Instruct-2507 - Qwen/Qwen3-4B-Thinking-2507 library_name: transformers datasets: - openai/gsm8k tags: - evoluation - math - merge --- # 📑 Model Card [💻Github Repo](https://github.com/Hert4/Evolution-Merge) • [🤗Model Collections](https://huggingface.co/collections/beyoru/evolution-model) ## Model Details This model is a merged version of two Qwen base models: - **Qwen/Qwen3-4B-Instruct-2507** - **Qwen/Qwen3-4B-Thinking-2507** ## Notations: - **Evoluation dataset**: `openai/gsm8k` (subset of 100 samples, not trained) - **Generation runs**: 50 - **Population size**: 10 - This model design for instruct model not reasoning model with same function like Qwen3-Instruct-2507 - **A good start for SFT or GRPO training.** ## Evaluation - For my evaluation in my agent benchmark is not surpass too much but only 3% with instruct model. - Surpass `openfree/Darwin-Qwen3-4B` (Evolution model) and base model in ACEBench. ```bibtex @misc{nafy_qwen_merge_2025, title = {Merged Qwen3 4B Instruct + Thinking Models}, author = {Beyoru}, year = {2025}, howpublished = {\url{https://huggingface.co/beyoru/EvolLLM}}, note = {Merged model combining instruction-tuned and reasoning Qwen3 variants.}, base_models = {Qwen/Qwen3-4B-Instruct-2507, Qwen/Qwen3-4B-Thinking-2507} }