You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Nov 1, 2024. It is now read-only.
I am an undergraduate student studying Deep Learning in Korea. I am impressed by your paper and doing experiments.
I thought gpu memory usage was only related to the parameters of the model, but using non-local blocks would use too much gpu memory.
Adding two non-local blocks to the model of the i3d inception structure, the total number of parameters increased by about 7%.
However, gpu memory usage has increased significantly. For example, in the i3d inception structure, six batches were available for each gpu, but adding two non-local blocks maximized two batches for each gpu.
Can you tell me if this is common or if I miscoded it?
(input size=224x224, time=64)
The text was updated successfully, but these errors were encountered:
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
I am an undergraduate student studying Deep Learning in Korea. I am impressed by your paper and doing experiments.
I thought gpu memory usage was only related to the parameters of the model, but using non-local blocks would use too much gpu memory.
Adding two non-local blocks to the model of the i3d inception structure, the total number of parameters increased by about 7%.
However, gpu memory usage has increased significantly. For example, in the i3d inception structure, six batches were available for each gpu, but adding two non-local blocks maximized two batches for each gpu.
Can you tell me if this is common or if I miscoded it?
(input size=224x224, time=64)
The text was updated successfully, but these errors were encountered: