Potentially, this is one of the best 8B models for RP if you find the right settings that overcome the occasional repetitions.

(I'll say in advance โ€” I'm not lucky with settings...)


All-Q3-8B-RP-0625

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the DARE TIES merge method using Qwen/Qwen3-8B-Base as a base.

Models Merged

The following models were included in the merge:

  • GreenerPastures/Bald-Beaver-8B
  • allura-org/remnant-qwen3-8b
  • allura-org/Q3-8B-Kintsugi

Configuration

The following YAML configuration was used to produce this model:

merge_method: dare_ties
base_model: Qwen/Qwen3-8B-Base
dtype: bfloat16
models:
  - model: GreenerPastures/Bald-Beaver-8B
    parameters:
      weight: 0.2
  - model: allura-org/Q3-8B-Kintsugi
    parameters:
      weight: 0.4
  - model: allura-org/remnant-qwen3-8b
    parameters:
      weight: 0.4
parameters:
  density: 0.35
Downloads last month
3
Safetensors
Model size
8.19B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Disya/All-Q3-8B-RP-0625