File size: 1,589 Bytes
890a33d c551c2f 890a33d c551c2f 890a33d c551c2f 890a33d c551c2f 890a33d c551c2f 890a33d c551c2f 890a33d c551c2f 890a33d c551c2f 890a33d c551c2f 890a33d c551c2f 890a33d c551c2f 890a33d c551c2f 890a33d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 |
---
base_model: meta-llama/Llama-3.1-8B-Instruct
library_name: peft
license: mit
datasets:
- imnim/multiclass-email-classification
language:
- en
tags:
- Email-classifier
- Email-labelling
- Fine-tuning
- peft
- lora
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
Model is finetuned for the task of email labelling. It labels the given email into one or more than one categories based on email subject and email body.
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
The model classifies emails into the following 10 categories: "Business", "Personal", "Promotions", "Customer Support", "Job Application",
"Finance & Bills", "Events & Invitations", "Travel & Bookings", "Reminders", "Newsletters"
I have prepared a synthetic but realistic dataset of 2,105 labeled emails. Each email includes a subject, body, and one or more categories.
- **Developed by:** imnim
- **Model type:** text-to-text
- **Language(s) (NLP):** English
- **Finetuned from model:** Llama-3.1-8B-Instruct
### Model Sources
<!-- Provide the basic links for the model. -->
- **Repository:** https://github.com/contributerMe/multi-label-email-classifier
- **Demo:** https://huggingface.co/spaces/imnim/Multi-labelEmailClassifier
## Technical Specifications
### Model Architecture and Objective
Auto-regressive language model that uses an optimized transformer architecture.
### Compute Infrastructure
Kaggle Notebook
#### Hardware
Trained on Kaggle's P100 GPU
### Framework versions
- PEFT 0.15.2 |