Skip to main content

Toy GPT next-token prediction using a 2-token context window.

Project description

Toy-GPT: train-300-context-2

PyPI version Latest Release Docs License: MIT CI Deploy-Docs Check Links Dependabot

Demonstrates, at very small scale, how a language model is trained.

This repository is part of a series of toy training repositories plus a companion client repository:

  • Training repositories produce pretrained artifacts (vocabulary, weights, metadata).
  • A web app loads the artifacts and provides an interactive prompt.

Contents

  • a small, declared text corpus
  • a tokenizer and vocabulary builder
  • a simple next-token prediction model
  • a repeatable training loop
  • committed, inspectable artifacts for downstream use

Scope

This is:

  • an intentionally inspectable training pipeline
  • a next-token predictor trained on an explicit corpus

This is not:

  • a production system
  • a full Transformer implementation
  • a chat interface
  • a claim of semantic understanding

Outputs

This repository produces and commits pretrained artifacts under artifacts/.

Training logs and evidence are written under outputs/ (for example, outputs/train_log.csv).

Quick start

See SETUP.md for full setup and workflow instructions.

Run the full training script:

uv run python src/toy_gpt_train/d_train.py

Run individually:

  • a/b/c are demos (can be run alone if desired)
  • d_train produces artifacts
  • e_infer consumes artifacts
uv run python src/toy_gpt_train/a_tokenizer.py
uv run python src/toy_gpt_train/b_vocab.py
uv run python src/toy_gpt_train/c_model.py
uv run python src/toy_gpt_train/d_train.py
uv run python src/toy_gpt_train/e_infer.py

Provenance and Purpose

The primary corpus used for training is declared in SE_MANIFEST.toml.

This repository commits pretrained artifacts so the client can run without retraining.

Annotations

ANNOTATIONS.md - REQ/WHY/OBS annotations used

Citation

CITATION.cff

License

MIT

SE Manifest

SE_MANIFEST.toml - project intent, scope, and role

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

toy_gpt_train_300_context_2-0.9.8.tar.gz (86.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

toy_gpt_train_300_context_2-0.9.8-py3-none-any.whl (26.3 kB view details)

Uploaded Python 3

File details

Details for the file toy_gpt_train_300_context_2-0.9.8.tar.gz.

File metadata

File hashes

Hashes for toy_gpt_train_300_context_2-0.9.8.tar.gz
Algorithm Hash digest
SHA256 cb11975dba42a5cb5e3f68dc5399abcac60a5b06549b89623b6f41b8ec9ec1d8
MD5 4866dd534e1a2ef3cf20b025c29cf436
BLAKE2b-256 b6eb56681ea98ad4a6c7b999469aa340f333d481aee182cc4428cd5796f36cb1

See more details on using hashes here.

Provenance

The following attestation bundles were made for toy_gpt_train_300_context_2-0.9.8.tar.gz:

Publisher: release-pypi-mkdocs-shared.yml on toy-gpt/train-300-context-2

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file toy_gpt_train_300_context_2-0.9.8-py3-none-any.whl.

File metadata

File hashes

Hashes for toy_gpt_train_300_context_2-0.9.8-py3-none-any.whl
Algorithm Hash digest
SHA256 8d5f787a071460c5291d7433e08943159f8772a01dc53cd4f5d6d0ec3f5b5462
MD5 f4e8acf5c3dc81c53b72f165a6b22988
BLAKE2b-256 c24336f99dc5e2eef91b80582db956dce9bebdfb8b78790ea1c59d185341cc8d

See more details on using hashes here.

Provenance

The following attestation bundles were made for toy_gpt_train_300_context_2-0.9.8-py3-none-any.whl:

Publisher: release-pypi-mkdocs-shared.yml on toy-gpt/train-300-context-2

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page