-
Notifications
You must be signed in to change notification settings - Fork 229
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
detach output from teacher #100
base: 0.x
Are you sure you want to change the base?
Conversation
Or we can detach the tensor in cwd loss (https://github.com/open-mmlab/mmrazor/blob/master/mmrazor/models/losses/cwd.py) , what do you think? |
I'm so sorry, I haven't found this PR due to my negligence. There are some users who will train both teacher and student at the same time. mmrazor/mmrazor/models/distillers/single_teacher.py Lines 172 to 176 in e4e9513
|
Would you like to reopen this pr and complete this feature together? |
yup it would be better. |
Codecov Report
@@ Coverage Diff @@
## master #100 +/- ##
==========================================
+ Coverage 56.58% 66.31% +9.72%
==========================================
Files 83 92 +9
Lines 2932 3369 +437
Branches 540 613 +73
==========================================
+ Hits 1659 2234 +575
+ Misses 1197 1033 -164
- Partials 76 102 +26
Flags with carried forward coverage won't be shown. Click here to find out more.
Continue to review full report at Codecov.
|
* align mmedit static cfg * add for test * update requirments * add dependencies from mmlab * change name * lower thresh for interrogate at first * update test * update to skip * Move import tensorrt * Move import statement Co-authored-by: SingleZombie <singlezombie@163.com>
When using Self Distiller with Channel Wise Distill, there would be a backward exception due to the tensor from teacher is not detached.
The PR fixed this bug.