-
Notifications
You must be signed in to change notification settings - Fork 3k
Update Relevance prompt to fix prompt, enhance evaluation steps and define output structure #41762
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
- Fix rubric 2 to not check "incorrect" instead check relevace only - Update output to get score and explanation only - Update prompt to be more crisp. Include evaluation steps
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Pull Request Overview
This pull request updates the relevance evaluator prompt to fix the relevance rubric, streamline the evaluation instructions, and clearly define the expected output format. The changes include refining the system and user instructions, updating the evaluation steps, and providing explicit sample output formats.
- Fixed the rubric to check relevance only.
- Enhanced and clarified evaluation steps.
- Defined a crisp output structure for score and explanation.
Comments suppressed due to low confidence (1)
sdk/evaluation/azure-ai-evaluation/azure/ai/evaluation/_evaluators/_relevance/relevance.prompty:152
- The sample output only includes and tags, whereas earlier instructions mention providing answers between , , and . Consider updating the sample outputs or the instructions to ensure consistency in the expected output format.
<S2>5</S2>
Relevance Evaluator Prompt Optimization SummaryPull Request: [#41762](#41762) Key Prompt Updates:
Evaluation Metrics:
Model-Wise Pairwise Agreement (%):
Summary of Improvements:
|
Looks good. May I ask you to add the mean and std for agreement metrics across models? |
sdk/evaluation/azure-ai-evaluation/azure/ai/evaluation/_evaluators/_relevance/relevance.prompty
Show resolved
Hide resolved
update examples for rubric 4 and 5
update examples for rubric 4 and 5
update examples for rubric 4 and 5
sdk/evaluation/azure-ai-evaluation/azure/ai/evaluation/_evaluators/_relevance/_relevance.py
Show resolved
Hide resolved
#sign-off |
…into ghyadav/relevance_v2_update # Conflicts: # sdk/evaluation/azure-ai-evaluation/azure/ai/evaluation/_evaluators/_relevance/_relevance.py
…into ghyadav/relevance_v2_update
Description
Relevance V1 Results:



Relevance V2 Results:



Relevance V2 [2025-07-02 update]:

Please add an informative description that covers that changes made by the pull request and link all relevant issues.
If an SDK is being regenerated based on a new swagger spec, a link to the pull request containing these swagger spec changes has been included above.
All SDK Contribution checklist:
General Guidelines and Best Practices
Testing Guidelines