eval_framework.metrics package

Subpackages

Submodules

eval_framework.metrics.base module

class eval_framework.metrics.base.BaseMetric[source]

Bases: ABC, Generic

KEYS: list[str] | None = None
NAME: str
NAMES
abstractmethod calculate(response)[source]
Return type:

list[MetricResult]

Parameters:

response (Response)

class eval_framework.metrics.base.MetricResult(**data)[source]

Bases: BaseModel

Parameters:
  • metric_name (str)

  • value (float | None)

  • higher_is_better (bool)

  • llm_judge_prompt (str | None)

  • llm_judge_response (str | None)

  • code_execution_trace (str | None)

  • error (Error | None)

code_execution_trace: str | None
error: Error | None
higher_is_better: bool
llm_judge_prompt: str | None
llm_judge_response: str | None
metric_name: str
model_config: ClassVar[ConfigDict] = {'extra': 'forbid'}

Configuration for the model, should be a dictionary conforming to [ConfigDict][pydantic.config.ConfigDict].

value: float | None
class eval_framework.metrics.base.classproperty(method)[source]

Bases: object

Parameters:

method (Any)

Module contents