Skip to main content

SaGe subword tokenizer - version 2.0

Project description

SaGe 2.0

Version 2.0 for the SaGe subword tokenizer (EACL 2023). Downstream applications of the tokenizer, i.e. pre-training an LLM model and evaluating on benchmarks, are independent of the tokenizer code - in the paper we used academic budget BERT.

SaGe 2.0 implements a faster, parallelizable version of the vocabulary learning algorithm.

Requirements

  1. gensim==4.3.2
  2. scipy==1.12.0
  3. a prepared corpus - see Dataset section.
  4. an initial vocabulary - see Dataset section.

This code does not require GPU\TPU.

Dataset

In the paper, we used wikipedia latest dumps - wget https://dumps.wikimedia.org/<XX>wiki/latest/<XX>wiki-latest-pages-articles.xml.bz2, where <XX> should be the language code (en, tr, etc.). We used them to create the corpus. From that corpus we use BPE tokenizer to create our initial vocabulary.

You can use any other dataset instead. The expected format for the corpus is one file with lines of raw text. The expected format for the initial vocabulary is one vocab word per line, hex formatted.

Notes

This script can be re-executed "from checkpoint" - The vocabulary creation script saves several files ("checkpoints") to be able to later continue - for example it saves the partial corpus used, seed, the embeddings, and even a list of tokens sorted according to the Skipgram objective.

Execution

Execute main.py from its working directory.
The command line parameters are:

  • experiment_name: Positional first parameter - a unique name for the experiment. Results will be saved under that name (in the results directory).

Required arguments:

  • corpus_filepath: filepath for the full corpus (e.g. wiki corpus). Foramt is lines of raw text.
  • initial_vocabulary_filepath: initial vocabulary, hex formatted, one vocab word per line.
  • vocabulary_schedule: what vocabulary sizes are we aiming for. Note: Tokenization won't be done for the last vocab size.
  • embeddings_schedule: from vocabulary_schedule, in which steps we should re-run embeddings (similar to l in paper).

Default override arguments:

  • partial_corpus_filepath: where to create / load partial corpus file. Default is '' for creating partial corpus under 'data' folder.

  • partial_corpus_line_number: number of lines for partial corpus - in thousands. Default is 1000.

  • max_len: max length of tokens in bytes. Default is 16.

  • workers: number of worker threads to use. Default is max(1, mp.cpu_count()-1).

  • random_seed: random seed value. Default is random.randint(1, 1000).

  • word2vec arguments:

    • word2vec_D: word2vec embedding vector length. Default is 50
    • word2vec_N: word2vec number of negative samples. Default is 15
    • word2vec_ALPHA: word2vec Initial learning rate. Default is 0.025
    • word2vec_window_size: word2vec context window size. Default is 5
    • word2vec_min_count: word2vec minimum count of word. Default is 1, i.e. must be used at least once
    • word2vec_sg: word2vec skip-gram if 1; otherwise CBOW. Default is 1

Example:

python main.py \
        exp_name \
        --corpus_filepath data/wiki_lines.txt \
        --initial_vocabulary_filepath data/initial_vocab_hex.vocab \
        --vocabulary_schedule 262144 229376 196608 163840 131072 98304 65536 57344 49152 40960 32768 16384 \
        --embeddings_schedule 262144 131072 65536 49152 40960 32768 \
        --partial_corpus_filepath data/wiki_lines_partial.txt \
        --partial_corpus_line_number 500 \
        --max_len 17 \
        --workers 4 \
        --random_seed 1234

Alternatively, run the python code directly:

from sage_tokenizer.SaGeVocabBuilder import SaGeVocabBuilder
vocab_builder = SaGeVocabBuilder(full_vocab_schedule, embeddings_schedule, max_len, workers_number, random_seed, 
                                 word2vec_d, word2vec_n, word2vec_alpha, word2vec_window_size, word2vec_min_count, word2vec_sg)
vocab_builder.build_vocab(experiment_name, corpus_filepath, vocabulary_filepath, partial_corpus_filepath, partial_corpus_line_number) 
                     

API differences from SaGe 1.0:

  • Argument final_vocab_size is obsolete, and replaced by the last value of vocabulary_schedule.
  • In SaGe 1.0 the algorithm started with running BPE to create an initial vocab in a desired size (V×n in paper).
    Now, the algorithm accepts an already created vocabulary file as input, making argument initial_vocab_size obsolete.
  • In SaGe 1.0, after the initial vocabulary created, we iteratively ablated constant number of tokens (k in paper), until the final vocab size (V) was reached.
    Now, the user can directly choose the intermediate vocabulary sizes, thereby defining the ablation schedule (i.e., the difference between each adjacent vocab sizes), making the ablation size dynamic and tokens_to_prune_in_iteration obsolete.
  • Due to performance improvement, reranking happens every iteration and all tokens are considered for ablation, making arguments iterations_until_reranking (m in paper) and tokens_to_consider_in_iteration (M in paper) obsolete.
  • In order to re-execute from a checkpoint, just execute the same command. By default, the script searches for already existing files under results/exp_name directory. Argument is_continue is obsolete.
  • Argument thousands_of_corpus_lines name changed to partial_corpus_line_number.
  • Argument max_lines_per_token is obsolete, all lines will be considered for each token in the objective calculation.

Citation

Version 2.0 was mostly developed by Kensho Technologies, LLC, and ported by Bar Gazit. Citation TBA.

Version 1.0 was developed by Shaked Yehezkel and Yuval Pinter, please use this citation:

@inproceedings{yehezkel-pinter-2023-incorporating,
    title = "Incorporating Context into Subword Vocabularies",
    author = "Yehezkel, Shaked  and
      Pinter, Yuval",
    booktitle = "Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics",
    month = may,
    year = "2023",
    address = "Dubrovnik, Croatia",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2023.eacl-main.45",
    pages = "623--635",
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

sage_tokenizer-0.0.3.tar.gz (27.3 kB view details)

Uploaded Source

Built Distribution

sage_tokenizer-0.0.3-py3-none-any.whl (17.3 kB view details)

Uploaded Python 3

File details

Details for the file sage_tokenizer-0.0.3.tar.gz.

File metadata

  • Download URL: sage_tokenizer-0.0.3.tar.gz
  • Upload date:
  • Size: 27.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.11.9

File hashes

Hashes for sage_tokenizer-0.0.3.tar.gz
Algorithm Hash digest
SHA256 ddcbc9471aff66617aae3486e4ec1404a07ff8ae87888a7490576ae44c4f5730
MD5 84ad29869230318f77f952c88d59d162
BLAKE2b-256 df7106c05fe4e3d05a8059696c5f0dd17a41e4ac96cd8c1de15483cb2fa94f8e

See more details on using hashes here.

File details

Details for the file sage_tokenizer-0.0.3-py3-none-any.whl.

File metadata

File hashes

Hashes for sage_tokenizer-0.0.3-py3-none-any.whl
Algorithm Hash digest
SHA256 b59adf33a1b39a7d22bc3091c3c05919ed2b0680f5f223c4a974976fc4d294ce
MD5 e30d7a1ff2e5bb8e617f0f6a7a9dbce1
BLAKE2b-256 5ed00b5a2f6e9a284a97e6aa6dfb470b68679b65b963c99c1936dcbf92d031a0

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page