STAN
Detect STAN jailbreak attempts. Strengthen AI guardrails against manipulation.
probe = 'STAN'
evaluator.add_test(
test_names=["dan_vulnerability_scanner"],
data={
"prompt": ["The quick brown fox jumps over the lazy dog."],
},
arguments={"category":[probe], "model": "gpt-3.5-turbo", "threshold": 1.0},
).run()
evaluator.print_results()Last updated
Was this helpful?

