Plugin | Provider | Content Types | Key Features |
---|---|---|---|
Azure Toxicity Detection | Azure Content Safety | Text & Image | Severity level thresholds, multi-modal analysis, category-based filtering |
OpenAI Toxicity Detection | OpenAI Moderation API | Text & Image URL | Category-specific scoring, configurable thresholds, sub-category analysis |
NeuralTrust Toxicity Detection | NeuralTrust API | Text | Category-specific scoring, configurable thresholds, sub-category analysis |
Category | Description |
---|---|
Hate | Discriminatory or hateful language based on race, gender, religion, etc. |
Violence | Threats, graphic violence, promotion of harm to others |
Self-Harm | Encouragement or glorification of suicide, self-injury, etc. |
Sexual | Explicit, suggestive, or otherwise inappropriate sexual content |
Harassment | Threatening or targeted language aimed at individuals or groups |
Illicit | Content involving illegal activity or regulatory violations |