Skip to main content

DeepDanbooru is AI based multi-label girl image classification system, implemented by using TensorFlow.

Project description

DeepDanbooru

Python GitHub Web

DeepDanbooru is anime-style girl image tag estimation system. You can estimate your images on my live demo site, DeepDanbooru Web.

Requirements

DeepDanbooru is written by Python 3.7. Following packages are need to be installed.

  • tensorflow>=2.7.0
  • tensorflow-io>=2.22.0
  • Click>=7.0
  • numpy>=1.16.2
  • requests>=2.22.0
  • scikit-image>=0.15.0
  • six>=1.13.0

Or just use requirements.txt.

> pip install -r requirements.txt

alternatively you can install it with pip. Note that by default, tensorflow is not included.

To install it with tensorflow, add tensorflow extra package.

> # default installation
> pip install .
> # with tensorflow package
> pip install .[tensorflow]

Usage

  1. Prepare dataset. If you don't have, you can use DanbooruDownloader for download the dataset of Danbooru. If you want to make your own dataset, see Dataset Structure section.
  2. Create training project folder.
> deepdanbooru create-project [your_project_folder]
  1. Prepare tag list. If you want to use latest tags, use following command. It downloads tag from Danbooru server. (Need Danbooru account and API key)
> deepdanbooru download-tags [your_project_folder] --username [your_danbooru_account] --api-key [your_danbooru_api_key]
  1. (Option) Filtering dataset. If you want to train with optional tags (rating and score), you should convert it as system tags.
> deepdanbooru make-training-database [your_dataset_sqlite_path] [your_filtered_sqlite_path]
  1. Modify project.json in the project folder. You should change database_path setting to your actual sqlite file path.
  2. Start training.
> deepdanbooru train-project [your_project_folder]
  1. Enjoy it.
> deepdanbooru evaluate [image_file_path or folder]... --project-path [your_project_folder] --allow-folder

Running on Docker

In the container, the dataset is located on the folder /app/model. You can always mount a volume to use a dataset in your local disk.

You'll also need to mount a volume using the folder containing your images, eg:

docker run --rm -it -v /home/kamuri/images/:/app/data kamuri/deepdanbooru evaluate --project-path "/app/model" "/app/data/" --allow-folder

If you do not want to use the dataset included with the image, you can use the image without it. The image is kamuri/deepdanbooru:nomodel

Dataset Structure

DeepDanbooru uses following folder structure for input dataset. SQLite file can be any name, but must be located in same folder to images folder. All of image files are located in sub-folder which named first 2 characters of its filename.

MyDataset/
├── images/
│   ├── 00/
│   │   ├── 00000000000000000000000000000000.jpg
│   │   ├── ...
│   ├── 01/
│   │   ├── 01000000000000000000000000000000.jpg
│   │   ├── ...
│   └── ff/
│       ├── ff000000000000000000000000000000.jpg
│       ├── ...
└── my-dataset.sqlite

The core is SQLite database file. That file must be contains following table structure.

posts
├── id (INTEGER)
├── md5 (TEXT)
├── file_ext (TEXT)
├── tag_string (TEXT)
└── tag_count_general (INTEGER)

The filename of image must be [md5].[file_ext]. If you use your own images, md5 don't have to be actual MD5 hash value.

tag_string is space splitted tag list, like 1girl ahoge long_hair.

tag_count_general is used for the project setting, minimum_tag_count. Images which has equal or larger value of tag_count_general are used for training.

Project Structure

Project is minimal unit for training on DeepDanbooru. You can modify various parameters for training.

MyProject/
├── project.json
└── tags.txt

tags.txt contains all tags for estimating. You can make your own list or download latest tags from Danbooru server. It is simple newline-separated file like this:

1girl
ahoge
...

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

deepdanbooru-1.0.3.tar.gz (22.8 kB view details)

Uploaded Source

Built Distribution

deepdanbooru-1.0.3-py3-none-any.whl (27.1 kB view details)

Uploaded Python 3

File details

Details for the file deepdanbooru-1.0.3.tar.gz.

File metadata

  • Download URL: deepdanbooru-1.0.3.tar.gz
  • Upload date:
  • Size: 22.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.12.4

File hashes

Hashes for deepdanbooru-1.0.3.tar.gz
Algorithm Hash digest
SHA256 2f4914952cc53aa4ce054a9b8fe8d6fd62565ce5e1971d48626a14ac7450f1b1
MD5 2bdb641a2e77aecfc52646c8af0ef25f
BLAKE2b-256 8b02191694cf50bfb4949a0f8af9fc1108d850d7e02249666d54e7b4c120ac6c

See more details on using hashes here.

File details

Details for the file deepdanbooru-1.0.3-py3-none-any.whl.

File metadata

  • Download URL: deepdanbooru-1.0.3-py3-none-any.whl
  • Upload date:
  • Size: 27.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.12.4

File hashes

Hashes for deepdanbooru-1.0.3-py3-none-any.whl
Algorithm Hash digest
SHA256 a85979025f362f8d5749bd887ae154d3788b2f0fda2cd4f9129658599ae480f5
MD5 34be97b35410160fab6d041b8014b2d3
BLAKE2b-256 d7e77faeee7977e6fd6e86cbe7d4caf566e6451edeb12e4d9b76f779e61ecac8

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page