TurkishReasoner-Llama3.1-8B

Model Description

TurkishReasoner-Llama8B leverages Meta's powerful Llama3.1-8B foundation model to deliver sophisticated reasoning capabilities in Turkish. Fine-tuned using GRPO techniques, this model excels at multistep reasoning processes with particular strength in mathematical problem-solving and logical deduction.

Key Features

  • Built on Meta's advanced Llama3.1-8B foundation
  • Optimized for Turkish reasoning tasks with structured output
  • Balanced performance-to-resource ratio (8B parameters)
  • Strong multilingual understanding with Turkish specialization
  • Trained using Group Relative Policy Optimization (GRPO)
  • Clear step-by-step reasoning with formatted solutions

Technical Specifications

  • Base Model: Meta/Llama3.1-8B
  • Parameters: 8 billion
  • Input: Text
  • Hardware Requirements: ~16GB VRAM
  • Training Infrastructure: NVIDIA Ada6000 GPU

Usage

This model is well-suited for a variety of Turkish reasoning applications:

  • Educational platforms requiring detailed explanations
  • Research tools analyzing complex problem-solving approaches
  • Development of Turkish-language assistants with robust reasoning
  • Applications requiring balanced performance and efficiency

Example Usage

from transformers import pipeline

pipe = pipeline("text-generation", model="Chan-Y/TurkishReasoner-Llama3.1-8B", device=0)

messages = [
    {"role": "system", "content": """Sen kullanıcıların isteklerine Türkçe cevap veren bir asistansın ve sana bir problem verildi.
Problem hakkında düşün ve çalışmanı göster.
Çalışmanı <start_working_out> ve <end_working_out> arasına yerleştir.
Sonra, çözümünü <SOLUTION> ve </SOLUTION> arasına yerleştir.
Lütfen SADECE Türkçe kullan."""},
    {"role": "user", "content": "121'in karekökü kaçtır?"},
]

response = pipe(messages)
print(response)

For more information or assistance with this model, please contact the developers:

Downloads last month
3
Safetensors
Model size
8.03B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support