gemma-3-12b-logql-finetune-transformer
Fine-tuned Gemma 3 12B for LogQL Query Generation [Standard HF]
This model is fine-tuned from Google's Gemma 3 12B to generate LogQL queries from natural language descriptions.
Model Information
- Base Model: google/gemma-3-12b-it
- Architecture: Gemma3
- Format: Standard Hugging Face / Transformers
- Use Case: LogQL query generation
Training Details
- Base model: google/gemma-3-12b-it
- Fine-tuned for LogQL query generation
- Trained on custom generated dataset (15000 samples, 5 epochs)
Usage with Hugging Face Transformers
This model can be loaded directly using the transformers library.
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model_name = "NiekWork/gemma-3-12b-logql-finetune-transformer"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16)
prompt = "find the 5xx errors from the last 1 hour"
input_ids = tokenizer(prompt, return_tensors="pt").input_ids
# Generate text
output = model.generate(input_ids, max_new_tokens=100)
print(tokenizer.decode(output[0], skip_special_tokens=True))
Files
This repository contains standard Hugging Face files:
config.json- Model configurationmodel.safetensorsorpytorch_model.bin- Model weightstokenizer.jsonandtokenizer_config.json- Tokenizer files
License
This model is released under the Apache 2.0 license.
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support