Learn how to evaluate Large Language Models for bias, toxicity, and quality using the custom-llm-eval Python package. This tool enables custom criteria defin...
Level: intermediate
By Unknown
Category: tools