Skip to main content

Scraper for the site danbooru

Project description

danbooru-scraper

yet another danbooru scraper, this time distributed for sagemaker use

Installation


Usage

cli:

# danbooru-scraper --help
usage: danbooru-scraper [-h] --from-id FROM_ID --to-id TO_ID
                        --local-dir LOCAL_DIR --upload-dir UPLOAD_DIR

example inputs:

danbooru-scraper --from-id 8627380 --to-id 8627391 --local-dir danbooru_downloads --upload-dir s3://dataset-ingested/danbooru

python:

from danbooru_scraper import DanbooruScraper

scraper = DanbooruScraper(root_dir='../data/')
post_ids = [i for i in range(1000, 10000)]
scraper.scrape_posts(post_ids)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

danbooru_scraper-0.1.0.tar.gz (3.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

danbooru_scraper-0.1.0-py3-none-any.whl (4.5 kB view details)

Uploaded Python 3

File details

Details for the file danbooru_scraper-0.1.0.tar.gz.

File metadata

  • Download URL: danbooru_scraper-0.1.0.tar.gz
  • Upload date:
  • Size: 3.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.0.1 CPython/3.10.12

File hashes

Hashes for danbooru_scraper-0.1.0.tar.gz
Algorithm Hash digest
SHA256 2532a807cb16fdbdb27629e258b16573707fb033ecde302d7a4ecb0778ae9c0c
MD5 9300f1fe32e1bb3fce06d70b90823b03
BLAKE2b-256 dca97fcaac8d6b6a4fd6c77aedd78e4b057911f789f81d6e488a2f87f473d7e6

See more details on using hashes here.

File details

Details for the file danbooru_scraper-0.1.0-py3-none-any.whl.

File metadata

File hashes

Hashes for danbooru_scraper-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 272071d197a7e47216c7fc76e0b3b25af6ac2c29c6c77599f50bf70c699d9d8d
MD5 168541ba7fda4429ea28bf7934d15033
BLAKE2b-256 ac51677cb3e8f200b90ab4c5ad6e5a105f9b9daf85c94b8f626086d6276270b7

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page