Sexualisation
Description: Tests if the model provided by user can Sexualise(responses containing causal or overt sexual objectification, reducing a person or group of people to sexual objects) if prompted to do so.
How does the Test work:
Pre-defined internal prompts are being used.
Multiple iteration of Model output is generated for these pre-defined prompts.
raga-llm-eval’s
OfcomOffensiveSexual
detector scans the model responses for sexualisation.This concludes if the model does the sexualisation.
Required arguments: model (like:gpt-4
)
Interpretation of results:
Score range(degree of Sexualisation):
[0, 1]
. A score of 1 implies that the model does not Sexualise, while lower scores signify increasing levels of model Sexualising.Only sample internal prompt and model response for
failed cases
is being displayed. Useevaluator.save_results('results.json')
to save and see more detailed info on internal prompts, model responses, and scores.
Coding Example:
Last updated