Skip to content

Support Flexible Inputs to vertexai.types.Metric #6261

@annedranowski

Description

@annedranowski

Many evaluation metrics do not require extra context provided by prompts. In fact, the prompt may confuse the metric. For example, a non_advice metric might start checking response on instruction-following confusing instructions for extensions to non_advice criteria.

My use case for evaluating responses without relying on a prompt is currently not supported.

Hardcoding an LLMMetric with only a {response} field, and

  • supplying an eval_dataset with only a response column raises EvalDatasetSchema.UNKNOWN
  • while supplying an eval_dataset with prompt and response columns raises INVALID_ARGUMENT

The current workarounds (supplying dummy prompt or extra 'ignore prompt' instructions to the judge) are cumbersome.

I propose adding infrastructure to support the bare minimum 'response' input, in the form of a more flexible evaluate -- so addressing the INVALID_ARGUMENT error.

This would be a welcome addition to the existing infrastructure supporting extra inputs.

Metadata

Metadata

Assignees

No one assigned

    Labels

    api: vertex-aiIssues related to the googleapis/python-aiplatform API.

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions