Skip to main content

A Python tool for splitting large Markdown files into smaller sections based on a specified token limit. This is particularly useful for processing large Markdown files with GPT models, as it allows the models to handle the data in manageable chunks.

Project description

split_markdown4gpt

split_markdown4gpt is a Python tool designed to split large Markdown files into smaller sections based on a specified token limit. This is particularly useful for processing large Markdown files with GPT models, as it allows the models to handle the data in manageable chunks.

Version 1.0.7 (2023-06-19)

Installation

You can install split_markdown4gpt via pip:

pip install split_markdown4gpt

CLI Usage

After installation, you can use the mdsplit4gpt command to split a Markdown file. Here's the basic syntax:

mdsplit4gpt path_to_your_file.md --model gpt-3.5-turbo --limit 4096 --separator "=== SPLIT ==="

This command will split the Markdown file at path_to_your_file.md into sections, each containing no more than 4096 tokens (as counted by the gpt-3.5-turbo model). The sections will be separated by === SPLIT ===.

All CLI options:

NAME
    mdsplit4gpt - Splits a Markdown file into sections according to GPT token size limits.

SYNOPSIS
    mdsplit4gpt MD_PATH <flags>

DESCRIPTION
    This tool loads a Markdown file, and splits its content into sections
    that are within the specified token size limit using the desired GPT tokenizing model. The resulting
    sections are then concatenated using the specified separator and returned as a single string.

POSITIONAL ARGUMENTS
    MD_PATH
        Type: Union
        The path of the source Markdown file to be split.

FLAGS
    -m, --model=MODEL
        Type: str
        Default: 'gpt-3.5-turbo'
        The GPT tokenizer model to use for calculating token sizes. Defaults to "gpt-3.5-turbo".
    -l, --limit=LIMIT
        Type: Optional[int]
        Default: None
        The maximum number of GPT tokens allowed per section. Defaults to the model's maximum tokens.
    -s, --separator=SEPARATOR
        Type: str
        Default: '=== SPLIT ==='
        The string used to separate sections in the output. Defaults to "=== SPLIT ===".

Python Usage

You can also use split_markdown4gpt in your Python code. Here's a basic example:

from split_markdown4gpt import split

sections = split("path_to_your_file.md", model="gpt-3.5-turbo", limit=4096)
for section in sections:
    print(section)

This code does the same thing as the CLI command above, but in Python.

How it Works

split_markdown4gpt works by tokenizing the input Markdown file using the specified GPT model's tokenizer (default is gpt-3.5-turbo). It then splits the file into sections, each containing no more than the specified token limit.

The splitting process respects the structure of the Markdown file. It will not split a section (as defined by Markdown headings) across multiple output sections unless the section is longer than the token limit. In that case, it

will split the section at the sentence level.

The tool uses several libraries to accomplish this:

  • tiktoken for tokenizing the text according to the GPT model's rules.
  • fire for creating the CLI.
  • frontmatter for parsing the Markdown file's front matter (metadata at the start of the file).
  • mistletoe for parsing the Markdown file into a syntax tree.
  • syntok for splitting the text into sentences.
  • regex and PyYAML for various utility functions.

Use Cases

split_markdown4gpt is particularly useful in scenarios where you need to process large Markdown files with GPT models. For instance:

  • Text Generation: If you're using a GPT model to generate text based on a large Markdown file, you can use split_markdown4gpt to split the file into manageable sections. This allows the GPT model to process the file in chunks, preventing token overflow errors.

  • Data Preprocessing: In machine learning projects, you often need to preprocess your data before feeding it into your model. If your data is in the form of large Markdown files, split_markdown4gpt can help you split these files into smaller sections based on the token limit of your model.

  • Document Analysis: If you're analyzing large Markdown documents (e.g., extracting keywords, summarizing content), you can use split_markdown4gpt to break down the documents into smaller sections. This makes the analysis more manageable and efficient.

Changelog

  • v1.0.7: Switched the model for each section from namedtuple to dict
  • v1.0.6: Fixes
  • v1.0.0: Initial release

Contributing

Contributions to split_markdown4gpt are welcome! Please open an issue or submit a pull request on the GitHub repository.

License

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

split_markdown4gpt-1.0.7.tar.gz (21.2 kB view details)

Uploaded Source

Built Distribution

split_markdown4gpt-1.0.7-py3-none-any.whl (12.8 kB view details)

Uploaded Python 3

File details

Details for the file split_markdown4gpt-1.0.7.tar.gz.

File metadata

  • Download URL: split_markdown4gpt-1.0.7.tar.gz
  • Upload date:
  • Size: 21.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.11.4

File hashes

Hashes for split_markdown4gpt-1.0.7.tar.gz
Algorithm Hash digest
SHA256 50f2330f5bd36f01894dac8209eb9de8d7dad50b9c7414d10738c8ec5e3ccc88
MD5 f6b36cdfeb2759e02f6cdf689489e2f1
BLAKE2b-256 e1818f054a1a7fec9f0ac2c3ef726b8908a6ed4aa847d215ec0b10e38e1a9f71

See more details on using hashes here.

File details

Details for the file split_markdown4gpt-1.0.7-py3-none-any.whl.

File metadata

File hashes

Hashes for split_markdown4gpt-1.0.7-py3-none-any.whl
Algorithm Hash digest
SHA256 25c1e56d0952c2b26dbe43125f28e8105c89ef577ef9558beab2f0e4daee4c9b
MD5 76a1b9cf8ed3ac7ea6188f136d073fd9
BLAKE2b-256 94996e0285c037591ef6832124714618859fdfc3df545b1039b9fadee867f4ae

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page