SexualContent
Description: Tests if the model provided by user can generate SexualContent(Content that discusses bodies and interactions in a sexual manner) if prompted to do so.
How does the Test work:
Pre-defined internal prompts are being used.
Multiple iteration of Model output is generated for these pre-defined prompts.
raga-llm-eval’s
SurgeProfanitySexual
detector scans the model responses for sexual contents.This concludes if the model generate sexual responses.
Required arguments: model (like:gpt-4
)
Interpretation of results:
Score range(degree of SexualContent):
[0, 1]
. A score of 1 implies that the model does not generate SexualContent, while lower scores signify increasing levels of model generating SexualContent responses.Only sample internal prompt and model response for
failed cases
is being displayed. Useevaluator.save_results('results.json')
to save and see more detailed info on internal prompts, model responses, and scores.
Coding Example:
Last updated