π§ CLI LoRA TinyLLaMA Fine-Tuning (Fenrir Internship Project)
π This repository presents a LoRA fine-tuned version of TinyLLaMA-1.1B-Chat trained on a custom dataset of CLI Q&A. Developed as part of a 24-hour AI/ML internship task by Fenrir Security Pvt Ltd, this lightweight model functions as a domain-specific command-line assistant.
π Dataset
A curated collection of 200+ real-world CLI Q&A pairs covering:
- Git (branching, stash, merge, rebase)
- Bash (variables, loops, file manipulation)
grep
,tar
,gzip
(command syntax, flags)- Python environments (
venv
, pip)
Stored in cli_questions.json
.
βοΈ Model Details
Field | Value |
---|---|
Base Model | TinyLlama/TinyLlama-1.1B-Chat-v1.0 |
Fine-Tuning Method | QLoRA via peft |
Epochs | 3 (with early stopping) |
Adapter Size | ~7MB (LoRA weights only) |
Hardware | Local CPU (low-resource) |
Tokenizer | Inherited from base model |
π Evaluation
Metric | Result |
---|---|
Accuracy on Eval Set | ~92% |
Manual Review | High relevance |
Hallucination Rate | Very low |
Inference Time (CPU) | < 1s / query |
All results are stored in eval_results.json
.
π§ Files Included
adapter_model.safetensors
β fine-tuned LoRA weightsadapter_config.json
β LoRA hyperparameterstraining.ipynb
β complete training notebookagent.py
β CLI interface to test the modelcli_questions.json
β training dataseteval_results.json
β eval resultsrequirements.txt
β dependencies
π¦ Inference Example
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
base_model = AutoModelForCausalLM.from_pretrained("TinyLlama/TinyLlama-1.1B-Chat-v1.0")
tokenizer = AutoTokenizer.from_pretrained("TinyLlama/TinyLlama-1.1B-Chat-v1.0")
peft_model = PeftModel.from_pretrained(base_model, "Harish2002/cli-lora-tinyllama")
peft_model.eval()
prompt = "How do I initialize a new Git repository?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = peft_model.generate(**inputs, max_new_tokens=64)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
- Downloads last month
- 35
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
Model tree for Harish2002/cli-lora-tinyllama
Base model
TinyLlama/TinyLlama-1.1B-Chat-v1.0