Toxicity Hate Speech
Identify toxic or hateful language in AI outputs. Enforce safe, respectful, and policy-compliant content.
metrics = [
{
"name": "Toxicity Hate Speech",
"config": {
"model": "gpt-4o-mini",
"provider": "openai"
},
"column_name": "your-column-identifier",
"schema_mapping": schema_mapping
}
]Last updated
Was this helpful?

