-
Notifications
You must be signed in to change notification settings - Fork 7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[RFC] New Augmentation techniques in Torchvison #3817
Comments
Would be nice to have these Augmentations in Also found this official code implementation for |
@oke-aditya Do we need the cutout given we have Random Erasing which can be configured to have the same more or less effect? |
I think the same. I compared both the implementations. RandomErasing is newer than Cutout, also both the augmentations produce almost similar results. Also, As per docs RandomErasing does not work for for PIL Images. It works only for |
Not a transform idea but what about adding an optional 'target_transforms' argument in class AugmentExample:
def __init__(self, p=0.1):
self.p = p
def __call__(self, img, box):
if self.p < torch.rand(1):
img = img.flip(-1)
box = box_hflip(box)
return img, box Also having a lot more keypoints/bbox transforms would be really great (ideally any image transform that involves a transformation of the targets should be accompanied by one?). (Sorry if my english isn't right, I speak baguette.) |
@datumbox is closing this issue intended? As I understand there is dataset and transforms rework. Which would be a major refactor Do we plan to migrate all transforms to new ones in near future? (I had a minor look at the proposal which looks fantastic) |
Not at all intended; Github just closed it when I merged the PR. We are certainly not done here :D |
@datumbox I'd like to take up ReMixMatch augmentation if no one's working on it. Would need some guidance on how to go around it though :) |
@lezwon Thanks a lot for offering to help! ReMixMatch focuses on learning augmentations and on using unlabelled data. One challenge with that is that the majority of the changes will have to land on references which are outside of TorchVision. Currently the reference scripts are in need of some rework to reduce the amount of duplicate code and improve the overall quality. It's on the top of our todos and until that's done, ideally we would like to avoid introducing significantly complex techniques like ReMixMatch. I wonder if you would be interested in implementing the AutoAugment Detection algorithm listed above. @vfdev-5 has already added most of the necessary low-level kernels for doing transforms on the BBoxes in BTW I'm currently working on the SimpleCopyPaste contribution trying to if we can train more accurate models using it. I'll let you know when I have the full results. :) |
@datumbox AutoAugment sounds good. I'll start looking into it. :) Also, I noticed your comment on SimpleCopyPaste PR. Lemme know if I can help in any way :) |
@lezwon Fantastic! Just note that I'm talking about this version designed for detection: AutoAugment Detection. This is different from the already supported algorithm for classification. :) |
Without talking about transforms implementation in the prototype (how input type dispatch would happen etc), the only thing I think we could miss to implement AA Detection is |
May I ask what's the current plan for the Fast AutoAugment? I have the implementation of another paper called Faster Autoaugment: Are maintainers interested in embedding this technique as well? If interested, what's the expected api to use it? Related issue: #5000 |
@ain-soph The Fast* AutoAugment methods are indeed on our radar. We should examine adding them after the work on the new Transforms API is complete. Let me explain why it's not primary target at this point:
One area we could use help is models and particularly Video architectures. Have a look at #2707 for some ideas. I hope the current situation won't discourage you from sticking around and continue contributing to TorchVision. We definitely want any help we can get from the community! :) |
🚀 Feature
Inclusion of new Augmentation techniques in
torchvision.transforms
.Motivation
Transforms are important for data augmentation 😅
Proposals
Cutout Citations 964Superseded by RandomErasingAdditional context
To visitors
Kindly give a 👍 if you think any of these would help in your work.
Also if you have any transform in mind please provide few details here!
Linked to #3221
cc @vfdev-5 @fmassa
The text was updated successfully, but these errors were encountered: