Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Incorporate bugbug ml labelling into moderation process #3574

Closed
ksy36 opened this issue May 19, 2021 · 6 comments
Closed

Incorporate bugbug ml labelling into moderation process #3574

ksy36 opened this issue May 19, 2021 · 6 comments

Comments

@ksy36
Copy link
Contributor

ksy36 commented May 19, 2021

In #3571 I've added bugbug classification and issues that are most likely invalid will receive bugbug-probability-high label in the private repository. We need to incorporate this into the moderation process to close such issues automatically.

There is a problem with posting issues content to the public repository without human moderation as we may end up submitting abusive content and get banned by GitHub again. So as a solution, we could post a placeholder text to the public issue, instead of its content for these invalid issues. Something like the current bot is posting now as a comment, but we'll put it in the issue body instead.

issue is labelled invalid by the bot -> do not post the content to the public repository, post a placeholder text instead and close the issue

If we decide to go this path, users won't be able to see the content of their report if it was labelled invalid (they will only see the placeholder text). It's not ideal (but acceptable?), and I wonder how important this is given the overwhelming amount of issues we're getting.

Do you have any thoughts on this?
@softvision-oana-arbuzov @softvision-raul-bucata @karlcow

@ksy36
Copy link
Contributor Author

ksy36 commented May 19, 2021

To add to this, we had a discussion with @karlcow about another approach. We could train a second model to detect nsfw sites, label those and make a decision about posting content to the public based on this label.

issue is labelled invalid and nsfw by the bot -> do not post the content to the public repository, post a placeholder text instead and close the issue

issue is labelled invalid by the bot -> post content to the public repository and close the issue

The problem with this approach is that we can't be 100% sure on nsfw prediction. We currently have a good amount of issues that are manually labelled nsfw, however, there always will be a new domain, an image or something on the page that could be considered abusive content, but will not be categorized by the model as nsfw. In addition, since we delete problematic (teen) issues in both repositories, there is no data for those domains either.

@softvision-oana-arbuzov
Copy link
Member

@ksy36 I think it is a good approach using the placeholder text, most of those issues have no value (meaning no relevant info is given to work with), and I don't think the users that report these kind of issues, will come back to comment.

@ksy36
Copy link
Contributor Author

ksy36 commented May 21, 2021

Sounds good, I'll go with this approach then, thanks for the feedback :)

@karlcow
Copy link
Member

karlcow commented May 31, 2021

Just deployed on staging.

I created a dumb issue which was identified as invalid and autoclosed and the original content was updated

This bug was identified as valid
and after switching to accepted the original content was published.

This is awesome @ksy36
This will accelerate the work of @softvision-oana-arbuzov and @softvision-raul-bucata

@ksy36
Copy link
Contributor Author

ksy36 commented May 31, 2021

thanks for testing on staging @karlcow, I'll deploy this on production today

@ksy36
Copy link
Contributor Author

ksy36 commented May 31, 2021

I think @karlcow has deployed it already 😁

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants