Skip to content

Conversation

sbite0138
Copy link

@sbite0138 sbite0138 commented Jul 18, 2025

The script provides a --max_examples option to limit the number of evaluation samples.
However, total is incremented and compared to max_examples before the per‑example score is recorded.
If --max_examples is specified, the statistics for the final sample are skipped, so the reported metrics are based on N – 1 examples instead of N.

This PR moves the total += 1 and the if max_examples and max_examples == total: break logic after the score update. This lets the last sample contribute to the metrics, ensuring the results accurately reflect all N examples.

@sbite0138 sbite0138 requested a review from a team as a code owner July 18, 2025 13:18
Copy link
Contributor

github-actions bot commented Jul 18, 2025

MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅

@sbite0138
Copy link
Author

recheck

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants