Replies: 1 comment
-
Speaking for my group, as soon as data augmentations were introduced, we stopped using feature caching. I can't remember if there was a slow-down during training, but the accuracy gain when using augmentations was worth it. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Feature caching is a functionality of our training code where pre-computed MFCC features for the training set will be saved to disk in-order, so that subsequent reads are faster. It was added to deal with a hardware setup for training that had slow IO which was bottlenecking the GPUs.
Since then, specially after the introduction of the augmentation features, feature caching has become a troublesome feature to support. We have documented the common pitfall of mixing augmentations and feature caching, but people still run into issues.
I'm opening this discussion to hopefully gather feedback from users of the training code: do you rely on feature caching? If so, why? What's your hardware setup and what benefit do you get from it? I'd love to understand more how others are using it so we can evaluate whether we want to keep this functionality or not.
Beta Was this translation helpful? Give feedback.
All reactions