-
Notifications
You must be signed in to change notification settings - Fork 171
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
accuracy metric #72
Comments
changed "output" to "answers" and it kind of fixed the problem but ... Processed prompts: 0%| | 0/10 [00:00<?, ?it/s]�[A Processed prompts: 10%|█ | 1/10 [00:00<00:03, 2.59it/s]�[A Processed prompts: 70%|███████ | 7/10 [00:00<00:00, 13.11it/s]�[A 19it [00:53, 2.81s/it] i think the code isnt complete yet... |
############################################################################################# modified the code and works |
Thx for the reply :D And as you pointed out the list out of index error, the author seems mistakenly put empty strings ("") in some gold answer lists in the eval_data they provide in this link. Not sure why, but this definitely doesn't look right and only lifts up the final metric score if using the |
The accuracy in metrics.py is defined as
While in run_short_form.py, acc is calculated per data instance:
where
pred
is some string, androw["output"]
is neither present in any short-form dataset, nor defined in your code.The text was updated successfully, but these errors were encountered: