enguard/tiny-guard-4m-en-general-politeness-multiclass-intel

This model is a fine-tuned Model2Vec classifier based on minishlab/potion-base-4m for the general-politeness-multiclass found in the Intel/polite-guard dataset.

Installation

pip install model2vec[inference]

Usage

from model2vec.inference import StaticModelPipeline

model = StaticModelPipeline.from_pretrained(
  "enguard/tiny-guard-4m-en-general-politeness-multiclass-intel"
)


# Supports single texts. Format input as a single text:
text = "Example sentence"

model.predict([text])
model.predict_proba([text])

Why should you use these models?

  • Optimized for precision to reduce false positives.
  • Extremely fast inference: up to x500 faster than SetFit.

This model variant

Below is a quick overview of the model variant and core metrics.

Field Value
Classifies general-politeness-multiclass
Base Model minishlab/potion-base-4m
Precision 0.9896
Recall 0.9783
F1 0.9839

Confusion Matrix

True \ Predicted impolite neutral polite somewhat polite
impolite 2477 31 5 19
neutral 13 2295 80 165
polite 3 91 2251 222
somewhat polite 10 218 245 2075
Full metrics (JSON)
{
  "impolite": {
    "precision": 0.9896124650419497,
    "recall": 0.9782780410742496,
    "f1-score": 0.9839126117179742,
    "support": 2532.0
  },
  "neutral": {
    "precision": 0.8709677419354839,
    "recall": 0.8989424206815512,
    "f1-score": 0.8847340015420201,
    "support": 2553.0
  },
  "polite": {
    "precision": 0.8721425803951957,
    "recall": 0.876899104012466,
    "f1-score": 0.8745143745143745,
    "support": 2567.0
  },
  "somewhat polite": {
    "precision": 0.8363563079403467,
    "recall": 0.8143642072213501,
    "f1-score": 0.8252137601908929,
    "support": 2548.0
  },
  "accuracy": 0.8919607843137255,
  "macro avg": {
    "precision": 0.8922697738282439,
    "recall": 0.8921209432474042,
    "f1-score": 0.8920936869913154,
    "support": 10200.0
  },
  "weighted avg": {
    "precision": 0.8920691454072528,
    "recall": 0.8919607843137255,
    "f1-score": 0.8919133038383807,
    "support": 10200.0
  }
}
Sample Predictions
Text True Label Predicted Label
I appreciate your interest in our vegetarian options. I can provide you with a list of our current dishes that cater to your dietary preferences. somewhat polite somewhat polite
I understand you're concerned about the ski lessons, and I'll look into the options for rescheduling. somewhat polite somewhat polite
Our technical skills course will cover the essential topics in data analysis, including data visualization and statistical modeling. The course materials will be available on our learning platform. neutral neutral
Our buffet hours are from 11 AM to 9 PM. Please note that we have a limited selection of options available during the lunch break. neutral neutral
I'll look into your policy details and see what options are available to you. somewhat polite somewhat polite
I appreciate your interest in our vegetarian options. I can provide you with a list of our current dishes that cater to your dietary preferences. somewhat polite somewhat polite
Prediction Speed Benchmarks
Dataset Size Time (seconds) Predictions/Second
1 0.0002 5096.36
1000 0.0551 18140.2
10000 0.6955 14377.94

Other model variants

Below is a general overview of the best-performing models for each dataset variant.

Classifies Model Precision Recall F1
general-politeness-binary enguard/tiny-guard-2m-en-general-politeness-binary-intel 0.9843 0.9889 0.9866
general-politeness-multiclass enguard/tiny-guard-2m-en-general-politeness-multiclass-intel 0.9875 0.9704 0.9789
general-politeness-binary enguard/tiny-guard-4m-en-general-politeness-binary-intel 0.9831 0.9878 0.9854
general-politeness-multiclass enguard/tiny-guard-4m-en-general-politeness-multiclass-intel 0.9896 0.9783 0.9839
general-politeness-binary enguard/tiny-guard-8m-en-general-politeness-binary-intel 0.9828 0.9905 0.9866
general-politeness-multiclass enguard/tiny-guard-8m-en-general-politeness-multiclass-intel 0.9873 0.9795 0.9833
general-politeness-binary enguard/small-guard-32m-en-general-politeness-binary-intel 0.9858 0.9889 0.9874
general-politeness-multiclass enguard/small-guard-32m-en-general-politeness-multiclass-intel 0.9897 0.9862 0.9879
general-politeness-binary enguard/medium-guard-128m-xx-general-politeness-binary-intel 0.9831 0.9901 0.9866
general-politeness-multiclass enguard/medium-guard-128m-xx-general-politeness-multiclass-intel 0.9881 0.9870 0.9876

Resources

Citation

If you use this model, please cite Model2Vec:

@software{minishlab2024model2vec,
  author       = {Stephan Tulkens and {van Dongen}, Thomas},
  title        = {Model2Vec: Fast State-of-the-Art Static Embeddings},
  year         = {2024},
  publisher    = {Zenodo},
  doi          = {10.5281/zenodo.17270888},
  url          = {https://github.com/MinishLab/model2vec},
  license      = {MIT}
}
Downloads last month
50
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train enguard/tiny-guard-4m-en-general-politeness-multiclass-intel

Collection including enguard/tiny-guard-4m-en-general-politeness-multiclass-intel