Skip to content

Bug Report – Test Cases Failing in Copilot Studio Kit #433

@gondor-netizen

Description

@gondor-netizen

Test Cases Failing with "AI-Generated Assessment to Review" in Copilot Studio Kit
Description:
We are encountering issues with test cases failing during Test Runs in Copilot Studio Kit. The following observations have been made:

  1. Test Failure Behavior:
    • Test cases with 'Generative Answers' selected and 'Answered' State often fail, resulting in a "2 - Failure" status.
    • The Result Reason field displays static text: "AI-generated Assessment of the Response. Please Review," but no detailed assessment is accessible.
  2. Generative Analysis Flow:
    • The Generative Analysis Flow executed during Test Runs provides only a cursory instruction to determine if the expected response is related to the question, assigning a simple status (Success, Failure, Unknown).
    • There is no mechanism to access or review the AI-generated assessment mentioned in the Result Reason field.
  3. Dataverse Table Limitations:
    • The Agent Test Result Table in Dataverse does not contain additional columns or data to provide insight into the AI-generated assessment or the reason for test failures.
      Request:
      We need assistance with the following:
  • Investigating why test cases are failing and whether detailed AI-generated assessments can be accessed or generated.
  • Identifying whether the Generative Analysis Flow or Dataverse Table configurations are contributing to the issue.
  • Providing guidance on resolving the "AI-generated Assessment to Review" issue and improving test result transparency.
    Priority: High

Metadata

Metadata

Labels

Type

Projects

Status

Backlog

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions