Replies: 1 comment
-
Greetings, @i3abghany! Is this effect compression-specific? Does it reproduce if you run training via the classification example without the |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello.
I am trying to generate multiple compressed models using NNCF. I am using CPU for training and I have access to so many cores. It seems to me that running the classification example with ResNet18 for CIFAR100 is best when run sequentially. When I run many instances using different configuration files, the performance is just terrible for individual tasks.
Why are runs interfering with each other and penalizing the overall performance? I am using the same dataset for all of them and I only maintain one copy of it. Could it be the case that training maintains locks over the files so only one run can have access to the data at any point in time?
I have tried different numbers of parallelly executed runs and it seems like it doesn't make much difference. Starting from 2 running processes, I can observe noticeable performance loss in individual runs.
Any leads will be appreciated. Thanks in advance!
Beta Was this translation helpful? Give feedback.
All reactions