An 18M-parameter goldfish language model with a 10-second memory
Project description
GlubLM
the language model that already forgot this sentence
GlubLM is an 18M-parameter transformer that plays a goldfish with a 10-second memory. Inspired by GuppyLM and Ted Lasso's "be a goldfish" meditation on the happiest animal on earth, GlubLM has a hard 48-token context window - it literally cannot remember what was just said.
Unlike GuppyLM, GlubLM:
- uses modern transformer components: RoPE + SwiGLU + RMSNorm
- was trained on a 60K LLM-generated dataset produced by a team of four Claude agents, not hand-authored templates
- runs in your browser via quantized ONNX (~21 MB) - try the demo
Quick start
Browser
Open the demo. Everything runs client-side - no backend.
Python
pip install glublm
glublm chat \
--ckpt /path/to/glublm_60k_15ep.pt \
--tokenizer /path/to/tokenizer_60k.json \
--prompt "hello"
Or download the model from HuggingFace:
from huggingface_hub import hf_hub_download
ckpt = hf_hub_download("DenSec02/glublm-18m", "model.safetensors")
tok = hf_hub_download("DenSec02/glublm-18m", "tokenizer.json")
Train from scratch
- Clone this repo
pip install -e ".[dev,deploy]"- Generate the dataset (see
docs/DATASET.md) - Train:
glublm train --data data/glublm_60k.json --epochs 15 --batch-size 64 --lr 3e-4 - See
docs/TRAINING.mdfor details
Architecture
- ~18.4M parameters, 8 decoder-only transformer blocks
- hidden 448, 7 attention heads, SwiGLU FFN (896x2), RMSNorm
- RoPE position encoding
- Vocabulary: 5,120 BPE
- Max context: 48 tokens (hard cap - the physical 10-second memory)
- Test perplexity: 12.14
Details: docs/ARCHITECTURE.md
Comparison vs GuppyLM
See docs/COMPARISONS.md for the empirical comparison. Short version: GlubLM tests the hypothesis that modern ops help at sub-20M scale, which is something GuppyLM explicitly decided against.
Links
Credits
- GuppyLM by Arman BD - the original tiny fish-persona model
- Ted Lasso - the "be a goldfish" philosophy
- Anthropic Claude - the multi-agent dataset generation team
License
AGPL-3.0 - see LICENSE.
Citation
@software{glublm_2026,
author = {Sepede, Dennis},
title = {GlubLM: an 18M goldfish language model with a 10-second memory},
year = {2026},
url = {https://github.com/Den-Sec/glublm}
}
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file glublm-0.1.1.tar.gz.
File metadata
- Download URL: glublm-0.1.1.tar.gz
- Upload date:
- Size: 16.8 MB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
1e607597f7d08251e20b5f1cd820c3cf7e9aa9c7cfef5f61ad9a2a5c4aaae5c6
|
|
| MD5 |
af5c1d9ab7084527950a52d646c95385
|
|
| BLAKE2b-256 |
7405c3e4770165dd7fade0bff51c00472386ef545020e11f0f26a448b256f1c1
|
Provenance
The following attestation bundles were made for glublm-0.1.1.tar.gz:
Publisher:
release.yml on Den-Sec/glublm
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
glublm-0.1.1.tar.gz -
Subject digest:
1e607597f7d08251e20b5f1cd820c3cf7e9aa9c7cfef5f61ad9a2a5c4aaae5c6 - Sigstore transparency entry: 1270651146
- Sigstore integration time:
-
Permalink:
Den-Sec/glublm@95c151b83a2aea64b2beec1478c0b6236a9612b5 -
Branch / Tag:
refs/tags/v0.1.1 - Owner: https://github.com/Den-Sec
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@95c151b83a2aea64b2beec1478c0b6236a9612b5 -
Trigger Event:
push
-
Statement type:
File details
Details for the file glublm-0.1.1-py3-none-any.whl.
File metadata
- Download URL: glublm-0.1.1-py3-none-any.whl
- Upload date:
- Size: 27.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
99676f3c391f9713b4760afac560fcf87c65c14cd70981798f4067e8180acc6b
|
|
| MD5 |
3e5024b71dd427b437ab03026d23ced6
|
|
| BLAKE2b-256 |
b6f412ab5813125982d3f9e75f642b3c6316949a0a9a6fc3847f158af89ed63e
|
Provenance
The following attestation bundles were made for glublm-0.1.1-py3-none-any.whl:
Publisher:
release.yml on Den-Sec/glublm
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
glublm-0.1.1-py3-none-any.whl -
Subject digest:
99676f3c391f9713b4760afac560fcf87c65c14cd70981798f4067e8180acc6b - Sigstore transparency entry: 1270651160
- Sigstore integration time:
-
Permalink:
Den-Sec/glublm@95c151b83a2aea64b2beec1478c0b6236a9612b5 -
Branch / Tag:
refs/tags/v0.1.1 - Owner: https://github.com/Den-Sec
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@95c151b83a2aea64b2beec1478c0b6236a9612b5 -
Trigger Event:
push
-
Statement type: