You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, Thanks for you excellent work on Dynamic Sparse Training. I am trying to reproduce your work and to reduce GPU memory consumption during sparse training.
I read through your code and try to implement your method on another vgg training implement to verify my understanding of your code. Here is my modification:
My run results show that ITOP with RigL consumes more or equal (it is supposed to be significantly less, right?) GPU memory than normal. Could you help me to figure out the problem of my implementation or any comment or suggestions?
Thanks.
The text was updated successfully, but these errors were encountered:
Hi, sorry for the late response. The sparse operation in this repo is implemented with masks due to the limited support of sparsity on GPU. Thus, the GPU memory consumption of ITOP will be weight memory + mask memory, which is larger than standard training.
Hi, Thanks for you excellent work on Dynamic Sparse Training. I am trying to reproduce your work and to reduce GPU memory consumption during sparse training.
I read through your code and try to implement your method on another vgg training implement to verify my understanding of your code. Here is my modification:
My modification:
bing0037/pytorch-vgg-cifar10_ITOP@6dab205
My running scripts:
Result: GPU memory consumption:
Baseline: 2639MB
ITOP with RigL: 2765MB
Question:
My run results show that ITOP with RigL consumes more or equal (it is supposed to be significantly less, right?) GPU memory than normal. Could you help me to figure out the problem of my implementation or any comment or suggestions?
Thanks.
The text was updated successfully, but these errors were encountered: