Skip to content

Conversation

mechakotik
Copy link
Contributor

Failed evaluation should never have better score than successful one.

Failed evaluation should never have better score than successful one.
@codelion
Copy link
Owner

This is not necessary as the evaluator.py can set the score to whatever needed for failed evaluations for a given example.

@mechakotik
Copy link
Contributor Author

It is useful when evaluator.py itself fails (syntax error, timeout, unhandled exception). Surely you can make an evaluator that takes care of all of this, but it's a nice fallback to not ruin all the progress if you didn't handle some edge case properly.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants