Skip to content

Code-based evaluators show score as Error in Azure Foundry evaluations #45643

@bastbu

Description

@bastbu
  • Package Name: azure-ai-projects
  • Package Version: 2.0.0
  • Operating System: Ubuntu
  • Python Version: 3.10

Describe the bug

There are a couple of issues with the ai-projects-sdk when running custom evaluators.

  • When running the sample for a custom code-based evaluator here, the UI in Foundry creates two metrics instead of one. The score is shown as "Error".
  • Logging output is not shown in the user logs, which makes it hard to debug custom scoring logic

To Reproduce

Steps to reproduce the behavior:

  1. Run the sample code

The my_custom_evaluator_code: score column shows an error:

Image

Note: the current sample code in the repository uses somewhat questionable queries, but I did not modify those for the purpose of running the sample in an unmodified manner.

Expected behavior

  • There should be an explanation of why this creates two metrics
  • The score should not display as "Error" when using the SDK
  • The sample should use neutral questions

Metadata

Metadata

Assignees

Labels

AI ProjectsService AttentionWorkflow: This issue is responsible by Azure service team.customer-reportedIssues that are reported by GitHub users external to the Azure organization.needs-team-attentionWorkflow: This issue needs attention from Azure service team or SDK teamquestionThe issue doesn't require a change to the product in order to be resolved. Most issues start as that

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions