-
Notifications
You must be signed in to change notification settings - Fork 109
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to manually check the testing result on each testing sample? #78
Comments
Hi, thanks for your attention on this project! For Q1, you can check the directory for data storage, and may need to transform the .pkl format into .json or .jsonl. For Q2, you need to clone the repo and add your own codes to implement this requirement. For example, you can add this when computing metrics or before with a for loop. Hope these can help you! Best regards |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Dear authors,
Thank you for sharing this awesome project!
I am trying to look deeper on the evaluation result. like, how exactly the model perform on different testing samples.
I successfully run over the example code you provided on README.md and got pretty good results. However, those results are limited to some high-level metrics.
So, I am trying to look deeper to the performance on each testing samples, to uncover some clues about:
Do you know how can I manually check the what the model is actually taking as input and output for each testing sample?
Thanks in advance!
Sincerely,
The text was updated successfully, but these errors were encountered: