Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Not accurate in real-world tests #2

Open
TheDailySpank opened this issue Mar 15, 2019 · 3 comments
Open

Not accurate in real-world tests #2

TheDailySpank opened this issue Mar 15, 2019 · 3 comments

Comments

@TheDailySpank
Copy link

I've been working on a similar idea for a while and then stumbled upon this via another project I've been following.

While the method works (images deemed inappropriate get blurred and the tag works) the classification is way off and is less than the 95% claimed.

For example, an image of a woman in a vineyard, fully clothed, with a wine bottle to her mouth was highly scored while images I won't describe but were very obviously pornographic went right through.

There is also bias towards grids of images, probably because of the thumbnail sheets of adult videos causing similar image layouts to be classified incorrectly. e.g. This image scores .52 on the explicit scale. Don't worry, it's totally safe. I've seen this issue with Yahoo's OPEN_NSFW release.

I have a large dataset of manually corrected images if you'd be willing to give me info on how to train from it. Initial classification was done using Yahoo's model.

@rstml
Copy link
Member

rstml commented Mar 15, 2019

@TheDailySpank thanks for the feedback! Indeed, it's far from being ready for real-world applications. 95% accuracy is only against the test data set which is limited.

One way to tune classification is to tweak cut-off point, e.g. set it to 0.75 instead of 0.5. It's currently not supported in this Chrome extension, but we plan to add it in the future releases.

Our goal is to engage the community and iteratively improve model robustness. To do this, we open sourcing as much as we can. You can use our training scripts which you can apply to your dataset. See this repo for more details: https://github.com/purify-ai/geacc-models

We are preparing a new version trained on 100K images (twice of current model) and plan to release the dataset itself.

We also accept data donations. If you can share your dataset, we could include it in our next release. However, due to legal and technical limitations, we can only accept data in the form of publicly accessible URLs.

@TheDailySpank
Copy link
Author

TheDailySpank commented Mar 15, 2019

I have a couple of databases with close to 5M images scored via OPEN_NSFW with ~100K "corrected" classifications, each about 50K. Meaning there's two piles, one of verified NSFW content and another with the corrections to the initial mis-classification.

@rstml
Copy link
Member

rstml commented May 24, 2019

@TheDailySpank we found an issue with the test dataset and re-tested. New results are much lower - 85% accuracy. Working on the next version which will be trained with more data and deeper model. So we should get better results.

Meanwhile, we're trying to get a better insight and I'd appreciate if you could take a few minutes to complete this survey https://forms.gle/6Wp9nUFeTXVgrNhB8

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants