Skip to content

categorizes and gives images a NSFW evaluation

Notifications You must be signed in to change notification settings

johnschriner/SFWpy

Folders and files

NameName
Last commit message
Last commit date

Latest commit

7e8f4a7 · May 17, 2017

History

3 Commits
May 17, 2017
May 17, 2017
May 17, 2017
May 17, 2017
May 17, 2017
May 17, 2017
May 17, 2017
May 17, 2017
May 17, 2017

Repository files navigation

SFWpy

categorizes and gives images a NSFW evaluation

input: url or image directory
output: categorized images and a NSFW evaluation and exports info to an SQLite database

SFWpy is very machine and environment-specific at the moment; it relies on paths for one particular machine and it's terribly dependent on environmental packages (i.e. it won't work on your machine without massive configuration)

For directories: the script reads in the images
For URLs: the script scrapes the site using ImageScraper and reads in the images.

Images are first processed by Inception in Tensorflow, a neural network for categorizing images with good accuracy. Then the images are processed through a Docker container supporting Caffe, the engine required for using Yahoo's open_nsfw evaluation.

Lastly, imagename/location as well as the categorization and NSFW evaluation are sent to an SQLite database

--To do:--
Flag images in db that score .7 or higher as NSFW for review
Add ability to change name of project
Work on interoperability -- may have to start over :/

About

categorizes and gives images a NSFW evaluation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages