Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to manually check the testing result on each testing sample? #78

Open
dayuyang1999 opened this issue Aug 18, 2023 · 1 comment
Open

Comments

@dayuyang1999
Copy link

dayuyang1999 commented Aug 18, 2023

Dear authors,

Thank you for sharing this awesome project!

I am trying to look deeper on the evaluation result. like, how exactly the model perform on different testing samples.

I successfully run over the example code you provided on README.md and got pretty good results. However, those results are limited to some high-level metrics.

So, I am trying to look deeper to the performance on each testing samples, to uncover some clues about:

  1. how the testing samples actually looks like to human?
  2. what is the performance of the model on each testing sample? and what are their recommended movies based on the historical dialog.

Do you know how can I manually check the what the model is actually taking as input and output for each testing sample?

Thanks in advance!

Sincerely,

@wxl1999
Copy link
Member

wxl1999 commented Aug 19, 2023

Hi, thanks for your attention on this project!

For Q1, you can check the directory for data storage, and may need to transform the .pkl format into .json or .jsonl.

For Q2, you need to clone the repo and add your own codes to implement this requirement. For example, you can add this when computing metrics or before with a for loop.

Hope these can help you!

Best regards

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants