image/png

Arcee-SuperNova-v1 (70B) is a merged model built from multiple advanced training approaches. At its core is a distilled version of Llama-3.1-405B-Instruct into Llama-3.1-70B-Instruct, using out DistillKit to preserve instruction-following strengths while reducing size.

Alongside this, another Llama-3.1-70B model was instruction-tuned using synthetic data from our Evol-Kit pipeline, improving precision and adherence across diverse queries. Updates were integrated mid-epoch for smoother performance gains.

A third version underwent Direct Preference Optimization (DPO) to better align with human feedback. While its contribution was smaller, it helped refine final alignment.

The resulting Arcee-SuperNova combines all three, delivering strong human preference alignment and state-of-the-art instruction-following ability.

Model Details

  • Architecture Base: Llama-3.1-70B-Instruct
  • Parameter Count: 70B
  • License: [Llama3]

Use Cases

  • General intelligence and instruction following
  • Serving as a base to be retrained over time using Reinforcement Learning from Human Feedback (RLHF)
  • Mathematical applications and queries

Quantizations

GGUF format available here

License

Arcee-SuperNova-v1 (70B) is released under the Llama-3 license. You are free to use, modify, and distribute this model in both commercial and non-commercial applications, subject to the terms and conditions of the license.

If you have questions or would like to share your experiences using Arcee-SuperNova-v1 (70B), please connect with us on social media. We’re excited to see what you build—and how this model helps you innovate!

Downloads last month
226
Safetensors
Model size
70.6B params
Tensor type
BF16
·
Inference Providers NEW
Input a message to start chatting with arcee-ai/Arcee-SuperNova-v1.

Model tree for arcee-ai/Arcee-SuperNova-v1

Finetuned
(78)
this model
Merges
1 model
Quantizations
3 models