You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Automate the evaluation process of model-generated solutions by comparing them against official solutions. This would streamline the validation and accuracy assessment. Yet, final human review would be still be necessary to avoid mistakes, but would be a much more faster phase for reviewers.
An simple and efficient process would be:
store answers in a subfolder of exam_path to be solved
during the solving process, at the end of each question, make a prompt asking to gpt-4o (cheaper compared to o1) to compare model's answer and offical answer and produce a concise result, much like langchain correctness prompts; save this comparison
when compiling the pdf, put the comparison after the model's answer for easy assessment
The text was updated successfully, but these errors were encountered:
Automate the evaluation process of model-generated solutions by comparing them against official solutions. This would streamline the validation and accuracy assessment. Yet, final human review would be still be necessary to avoid mistakes, but would be a much more faster phase for reviewers.
An simple and efficient process would be:
exam_path
to be solvedThe text was updated successfully, but these errors were encountered: