Skip to main content

A utility to extract vocabulary lists from manga.

Project description

Japanese Vocabulary Extractor

This script allows you to automatically scan through various types of japanese media and generate a CSV with all contained words for studying. Currently supported formats are:

  • Manga (as images)
  • Subtitles (ASS/SRT files) from anime, shows or movies
  • PDF and EPUB files
  • Text (txt) files

It allows you to automatically add the english definitions of each word to the CSV, as well as furigana if desired. It also allows creating individual CSVs for each file/volume and creates another large CSV with all vocab divided by file/volume (useful for Bunpro Sections).

The resulting CSV can be imported to Anki (if you add the english definitions) or Bunpro.

Installation

You need to have Python installed on your computer. I recommend using Python 3.12.

To install the Japanese Vocabulary Extractor, follow these steps:

  1. Open a terminal or command prompt on your computer.
  2. Type the following command and press Enter:
    pip install japanese-vocabulary-extractor
    

This will download and install the necessary files for the tool to work.

Usage

To use the Japanese Vocabulary Extractor, follow these steps:

  1. Open a terminal or command prompt on your computer.
  2. Type the following command and press Enter:
    jpvocab-extractor --type TYPE input_path
    

Replace TYPE with the type of media you are scanning: 'manga', 'subtitle', 'pdf', 'epub', 'txt' or 'generic'.

Replace input_path:

  • For manga, provide a folder containing the images.
  • For other types, provide the file or a folder with multiple files. Use quotation marks if the path has spaces.

This will create a vocab_all.csv file with all the words found.

Options

You can add options to the command to change its behavior. For example: To add English definitions to the CSV, include the --add-english option:

jpvocab-extractor --add-english --type TYPE input_path

Here is a list of all options:

  • --add-english: Looks up and adds the English translation of each word to the CSV file.
  • --furigana: Add furigana to all words in the CSV file. Note that this is quite primitive, it just adds the reading of the whole word in hiragana in brackets.
  • --id: Replaces each word with its JMDict ID in the CSV file. Incompatible with the --furigana flag.
  • --separate: Each volume/file will be saved to a separate CSV file. This also created one big combined vocab_combined CSV file with all vocab for each file/chapter in its own section, with duplicates removed. Requires --parent for manga.

There is one option only used for manga:

  • --parent: Only relevant if processing a manga: provided folder contains multiple volumes. Each folder will be treated as its own volume.

Here are all the available options shown together:

jpvocab-extractor [-h] [--parent] [--separate] [--id] [--add-english] [--furigana] --type TYPE input_path

Bunpro

The setup you'd want for Bunpro isn't known yet, but I'll put the expected command that should work best for manga here:

jpvocab-extractor --parent --separate --id --type manga input_path

This would combine all volumes into one CSV file, with JMDict IDs for each word. Each section will then correspond to one volume.

For general creation of decks for media other than manga, you would only add the --separate and --id flag:

jpvocab-extractor --separate --id --type TYPE input_path

This will separate all vocab into sections for each file within the CSV. If you do not have multiple files or the need for sections in your deck, leave out --separate.

Mokuro files

Bonus: Using this script with manga will also generate .mokuro and .html files for each volume, allowing you to read the manga with selectable text in your browser. For more details, visit the mokuro GitHub page linked at the bottom.

Notices

If you run into errors, look into the mokuro repository linked at the bottom. There might be some issues with python version compatibility.

Also important: This script is not perfect. The text recognition can make mistakes and some of the extracted vocab can be wrong. If this proves to be a big issue I will look for a different method to parse vocabulary from the text. Do not be alarmed by the warning about words with no definition, these are likely names, hallucinations/mistakes by the OCR algorithm or chinese symbols (sometimes found in subtitles).

TODO

  • Separate files/csv sections for each file for formats other than manga
  • Better furigana after each kanji instead of just the whole word
  • More advanced dictionary lookup functionality
  • Support more input formats (Games, VNs, Audio files?) Please suggest any you might want, even the ones listed already!
  • Support other output formats
  • Improve dictionary result accuracy to include one-character-kana words when translating to english (currently filtered out due to mostly useless answers)

Acknowledgements

This is hardly my work, I just stringed together some amazing libraries:

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

japanese_vocabulary_extractor-1.1.0.tar.gz (21.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

japanese_vocabulary_extractor-1.1.0-py3-none-any.whl (24.5 kB view details)

Uploaded Python 3

File details

Details for the file japanese_vocabulary_extractor-1.1.0.tar.gz.

File metadata

File hashes

Hashes for japanese_vocabulary_extractor-1.1.0.tar.gz
Algorithm Hash digest
SHA256 f5516dfc53a57c606b1438cd2c047bc4cd37fbb2cccc2ee887feb287c5cc8337
MD5 315daac0c78717bf5649009c34c3a5c4
BLAKE2b-256 e0f81ffeb203e7ee4dcb9f72f97cbde9f832e731c2a4f8d744135d31e5735ecc

See more details on using hashes here.

File details

Details for the file japanese_vocabulary_extractor-1.1.0-py3-none-any.whl.

File metadata

File hashes

Hashes for japanese_vocabulary_extractor-1.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 dd42aa0a5bb1b847fe3a57bd83ee9c5b0619d95e9194e57e03352b8266c86724
MD5 4c99e4489d2534d40479349d28b7c792
BLAKE2b-256 b1fadd8fdc8aad83bd9997063a6fbb86c46d75ec8d13accf2f5309cd181fcbbf

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page