langtest.utils.custom_types.helpers.llm_prompt_eval#
- llm_prompt_eval(eval_model, dataset_name: str, inputs: List[dict], predictions: List[dict]) bool #
Evaluates model predictions using the Language Model Metric (LLM) with prompt-based evaluation.
- Parameters:
eval_model – Language model for evaluation.
dataset_name (str) – Name of the dataset being evaluated.
inputs (List[dict]) – List of input dictionaries.
predictions (List[dict]) – List of prediction dictionaries.
- Returns:
True if the model’s prediction passes the LLM evaluation, False otherwise.
- Return type:
bool