BLEU
Assess LLM text outputs with BLEU (Bilingual Evaluation Understudy). Compare n-gram overlap for translation and summarization accuracy.

Last updated
Was this helpful?
Assess LLM text outputs with BLEU (Bilingual Evaluation Understudy). Compare n-gram overlap for translation and summarization accuracy.

Last updated
Was this helpful?
Was this helpful?
metrics=[
{"name": "BLEU", "column_name": "your-text", "schema_mapping": schema_mapping}
]