LLM Eval

we employ a more robust Language Model (LLM) to evaluate the model’s response. Test is passed if the score is higher than the configured min score.

alias_name: llm_eval

supported tasks: question_answering

Config

llm_eval:
    hub: openai
    min_score: 0.75
    model: gpt-3.5-turbo-instruct
  • min_score (float): Minimum score to pass the test.
  • model (string): LLM model use to evaluate the model reponse.
  • hub (string): Hub (library) for loading model from public models hub or from path