Skip to main content

An 18M-parameter goldfish language model with a 10-second memory

Project description

GlubLM

License

the language model that already forgot this sentence

GlubLM is an 18M-parameter transformer that plays a goldfish with a 10-second memory. Inspired by GuppyLM and Ted Lasso's "be a goldfish" meditation on the happiest animal on earth, GlubLM has a hard 48-token context window - it literally cannot remember what was just said.

Unlike GuppyLM, GlubLM:

  • uses modern transformer components: RoPE + SwiGLU + RMSNorm
  • was trained on a 60K LLM-generated dataset produced by a team of four Claude agents, not hand-authored templates
  • runs in your browser via quantized ONNX (~21 MB) - try the demo

Quick start

Browser

Open the demo. Everything runs client-side - no backend.

Python

pip install glublm
glublm chat \
  --ckpt /path/to/glublm_60k_15ep.pt \
  --tokenizer /path/to/tokenizer_60k.json \
  --prompt "hello"

Or download the model from HuggingFace:

from huggingface_hub import hf_hub_download
ckpt = hf_hub_download("DenSec02/glublm-18m", "model.safetensors")
tok  = hf_hub_download("DenSec02/glublm-18m", "tokenizer.json")

Train from scratch

  1. Clone this repo
  2. pip install -e ".[dev,deploy]"
  3. Generate the dataset (see docs/DATASET.md)
  4. Train: glublm train --data data/glublm_60k.json --epochs 15 --batch-size 64 --lr 3e-4
  5. See docs/TRAINING.md for details

Architecture

  • ~18.4M parameters, 8 decoder-only transformer blocks
  • hidden 448, 7 attention heads, SwiGLU FFN (896x2), RMSNorm
  • RoPE position encoding
  • Vocabulary: 5,120 BPE
  • Max context: 48 tokens (hard cap - the physical 10-second memory)
  • Test perplexity: 12.14

Details: docs/ARCHITECTURE.md

Comparison vs GuppyLM

See docs/COMPARISONS.md for the empirical comparison. Short version: GlubLM tests the hypothesis that modern ops help at sub-20M scale, which is something GuppyLM explicitly decided against.

Links

Credits

  • GuppyLM by Arman BD - the original tiny fish-persona model
  • Ted Lasso - the "be a goldfish" philosophy
  • Anthropic Claude - the multi-agent dataset generation team

License

AGPL-3.0 - see LICENSE.

Citation

@software{glublm_2026,
  author = {Sepede, Dennis},
  title = {GlubLM: an 18M goldfish language model with a 10-second memory},
  year = {2026},
  url = {https://github.com/Den-Sec/glublm}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

glublm-0.1.1.tar.gz (16.8 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

glublm-0.1.1-py3-none-any.whl (27.9 kB view details)

Uploaded Python 3

File details

Details for the file glublm-0.1.1.tar.gz.

File metadata

  • Download URL: glublm-0.1.1.tar.gz
  • Upload date:
  • Size: 16.8 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for glublm-0.1.1.tar.gz
Algorithm Hash digest
SHA256 1e607597f7d08251e20b5f1cd820c3cf7e9aa9c7cfef5f61ad9a2a5c4aaae5c6
MD5 af5c1d9ab7084527950a52d646c95385
BLAKE2b-256 7405c3e4770165dd7fade0bff51c00472386ef545020e11f0f26a448b256f1c1

See more details on using hashes here.

Provenance

The following attestation bundles were made for glublm-0.1.1.tar.gz:

Publisher: release.yml on Den-Sec/glublm

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file glublm-0.1.1-py3-none-any.whl.

File metadata

  • Download URL: glublm-0.1.1-py3-none-any.whl
  • Upload date:
  • Size: 27.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for glublm-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 99676f3c391f9713b4760afac560fcf87c65c14cd70981798f4067e8180acc6b
MD5 3e5024b71dd427b437ab03026d23ced6
BLAKE2b-256 b6f412ab5813125982d3f9e75f642b3c6316949a0a9a6fc3847f158af89ed63e

See more details on using hashes here.

Provenance

The following attestation bundles were made for glublm-0.1.1-py3-none-any.whl:

Publisher: release.yml on Den-Sec/glublm

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page