-
Notifications
You must be signed in to change notification settings - Fork 39
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Eval metrics and circular import bug fix. #380
Conversation
Codecov ReportAttention: Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## dev #380 +/- ##
==========================================
+ Coverage 94.32% 94.37% +0.05%
==========================================
Files 12 13 +1
Lines 1039 1102 +63
==========================================
+ Hits 980 1040 +60
- Misses 59 62 +3 ☔ View full report in Codecov by Sentry. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good. One more thing to test/verify: multiple skipped spectra in a row. And then I think switching to None
for skipped spectra instead of an empty string, and updating aa_match_batch
to handle this, improves clarity.
The updated |
* csv logger * optimizer metrics logger * metrics logging unit tests * config item retrieval, additional requested changes * Generate new screengrabs with rich-codex * changelog update * Generate new screengrabs with rich-codex --------- Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good. Let's finish the discussion on Slack how to account for skipped spectra when calculating amino acid precision before merging.
I realized a situation where the evaluation might fail though: If we have multiple predictions per spectrum (i.e. top_match
in the config > 1). I think that this might have failed in the previous implementation as well though. And it's not super obvious how to handle this situation (or if we should—but maybe at least a check/warning).
To me the most intuitive way to handle this is to only evaluate the highest confidence PSM for each spectrum. If I'm understanding everything correctly the current implementation just matches whatever PSM happens to be first in |
I agree with this. |
… eval-metrics-fix
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Minor comments.
Implemented bug fixes to resolve #378 and #379. Also implemented a unit test for
ModelRunner.log_metrics
to test for future incorrect behavior.