Skip to main content

Crawl your personal favorite images, photo albums, comics from website. Support pixiv, yande.re for now.

Project description

FavoritesCrawler

Build Version

Crawl your personal favorite images, photo albums, comics from website.

Warning!

  • Not ready for production.
  • Appropriately reduce the crawling speed in the future and provide options to set performance, but your account is still at risk of being disabled by the website.

Plan to support

  • instagram.com

Already support

  • pixiv.net (crawl your liked illust, must login), Thanks for project PixivPy.
  • yande.re (crawl your voted posts, require your username)
  • lmmpic.com (crawl your favorite albums, must login)
  • nhentai.net (crawl your favorite comics, must login)
  • twitter.com (crawl your liked posts, must login)

Requirements

  • Python3.7+

Install

pip install -U favorites_crawler

Config Proxy (Optional)

# on Windows
set https_proxy=http://localhost:8080  # replace with your proxy server
# on Liunx/macOS
export https_proxy=http://localhost:8080

Login

favors login [-h] {pixiv,yandere}

Login Pixiv

Thanks for @ZipFile Pixiv OAuth Flow

  1. run command
    favors login pixiv
    
  2. input your user_id (Access your pixiv personal page, copy from address bar), after press Enter, Pixiv login page will open in browser.
  3. Open dev console (F12) and switch to network tab.
  4. Enable persistent logging ("Preserve log").
  5. Type into the filter field: callback?
  6. Proceed with Pixiv login.
  7. After logging in you should see a blank page and request that looks like this: https://app-api.pixiv.net/web/v1/users/auth/pixiv/callback?state=...&code=.... Copy value of the code param into the prompt and hit the Enter key.

Login Yandere

  1. run command:
    favors login yandere
    
  2. input your username and hit the Enter key.

Login Lmmpic

  1. Open lmmpic on browser and login.
  2. Use "Get cookies.txt" extension download cookie file.
  3. Copy cookie file to {user_home}/.favorites_crawler.

Login NHentai

  1. Open nhentai on browser and login.
  2. Use "Get cookies.txt" extension download cookie file.
  3. Copy cookie file to {user_home}/.favorites_crawler.

Login Twitter

  1. run command
    favors login twitter
    
  2. input your username, after press Enter, likes page will open in browser.
  3. Open dev console (F12) and switch to network tab.
  4. Enable persistent logging ("Preserve log").
  5. Type into the filter field: Likes?
  6. Refresh Page.
  7. Copy Authorization, X-Csrf-Token and RequestURL from request(Likes?variables...) input on terminal.
  8. Use "Get cookies.txt" extension download cookie file.
  9. Copy cookie file to {user_home}/.favorites_crawler.

Crawl

favors crawl [-h] {lemon,nhentai,pixiv,yandere}

Crawl Pixiv

Before run this command, make sure you are already login.

favors crawl pixiv

Crawl Yandere

Before run this command, make sure you are already login.

favors crawl yandere

Crawl Lmmpic

Before run this command, make sure you are already login.

favors crawl lemon

Crawl NHentai

Before run this command, make sure you are already login.

favors crawl nhantai

Crawl Twitter

Before run this command, make sure you are already login.

favors crawl twitter

Config

Config file locate on {your_home}/.favorites_crawler/config.yml. You can set any scrapy built-in settings in this file.

By default, file content likes this:

pixiv:
  ACCESS_TOKEN: xxxxxxxxxxxxxxxxxxxxxxxxxxxx
  REFRESH_TOKEN: xxxxxxxxxxxxxxxxxxxxxxxxxxxx
  USER_ID: xxxx
yandere:
  USERNAME: xxxx

Download location

By default, pictures will download to working directory.
If you want to change download location, you can add FILES_STORE option to config.
For example, if you want save pixiv files to pictures/a, and want save yandere files to pictures/b, you can modify config file like this:

pixiv:
  ACCESS_TOKEN: xxxxxxxxxxxxxxxxxxxxxxxxxxxx
  REFRESH_TOKEN: xxxxxxxxxxxxxxxxxxxxxxxxxxxx
  USER_ID: xxxx
  FILES_STORE: pictures/a
yandere:
  USERNAME: xxxx
  FILES_STORE: pictures/b

Organize file by artist

if you want to organize pixiv illust by user, add this line to your config:

pixiv:
  # FAVORS_PIXIV_ENABLE_ORGANIZE_BY_USER: true  # (Deprecation)
  ENABLE_ORGANIZE_BY_ARTIST: true  # add this line to your yandere config

if you want to organize yandere post by artist, add this line to your config:

yandere:
  ENABLE_ORGANIZE_BY_ARTIST: true  # add this line to your yandere config

Store tags to IPTC/Keywords

only support pixiv and yandere.

yandere:
   ENABLE_WRITE_IPTC_KEYWORDS: true  # default: true
   EXIF_TOOL_EXECUTABLE: '<Path to your exiftool executable>'  # default None
pixiv:
   ENABLE_WRITE_IPTC_KEYWORDS: true  # default: true
   EXIF_TOOL_EXECUTABLE: '<Path to your exiftool executable>'  # default None

Restore your favorites

Vote yandere posts

$ favors restore yandere -h
usage: favors restore yandere [-h] -s {0,1,2,3} -t CSRF_TOKEN -c COOKIE path

positional arguments:
  path                  The location of the post to vote. (Sub-folders are ignored)

optional arguments:
  -h, --help            show this help message and exit
  -s {0,1,2,3}, --score {0,1,2,3}
                        Set 1, 2 or 3 to vote, 0 to cancel vote.
  -t CSRF_TOKEN, --csrf-token CSRF_TOKEN
                        CSRF token. To get it: 
                        1. Open your browser DevTools. 
                        2. Switch to network tab. 
                        3. Vote any post on yandere. 
                        4. Copy x-csrf-token value from request headers.
  -c COOKIE, --cookie COOKIE
                        Cookie. To get it: 
                        1. Open your browser DevTools. 
                        2. Switch to network tab. 
                        3. Vote any post on yandere. 
                        4. Copy cookie value from request headers.

Example:

favors restore yandere -s 3 -t "xxxx" -c "xx=x; xx=x; xx=x" .

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

favorites_crawler-0.2.0.tar.gz (24.1 kB view hashes)

Uploaded Source

Built Distribution

favorites_crawler-0.2.0-py3-none-any.whl (26.0 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page