Rule-based prose linter for formulaic AI writing patterns.
Project description
slop-guard
A rule-based prose linter that scores text 0--100 for formulaic AI writing patterns. No LLM judge, no API calls. Purely programmatic.
It runs ~80 compiled patterns against your text and returns a numeric score, a list of specific violations with surrounding context, and concrete advice for each hit.
Add to Your Agent
Claude Code
Add from the command line:
claude mcp add slop-guard -- uvx slop-guard
# Optional custom rule config:
claude mcp add slop-guard -- uvx slop-guard -c /path/to/config.jsonl
Add to your .mcp.json:
{
"mcpServers": {
"slop-guard": {
"command": "uvx",
"args": ["slop-guard"]
}
}
}
With a custom rule config:
{
"mcpServers": {
"slop-guard": {
"command": "uvx",
"args": ["slop-guard", "-c", "/path/to/config.jsonl"]
}
}
}
Codex
Add from the command line:
codex mcp add slop-guard -- uvx slop-guard
# Optional custom rule config:
codex mcp add slop-guard -- uvx slop-guard -c /path/to/config.jsonl
Add to your ~/.codex/config.toml:
[mcp_servers.slop-guard]
command = "uvx"
args = ["slop-guard"]
With a custom rule config:
[mcp_servers.slop-guard]
command = "uvx"
args = ["slop-guard", "-c", "/path/to/config.jsonl"]
If you want a fixed release, pin it in args, for example: ["slop-guard==0.3.0"].
CLI
The sg command lints prose files from the terminal. No API keys, no network calls.
Quick start
# Run without installing
uvx --from slop-guard sg README.md
# Or install it
uv tool install slop-guard
sg README.md
Usage
sg [OPTIONS] INPUT [INPUT ...]
sg requires at least one input. Each input can be a file path, - for stdin, or quoted inline prose text:
sg "This is some test text"
echo "This is a crucial paradigm shift." | sg -
Lint multiple files at once (shell-level glob expansion):
sg docs/*.md README.md
sg path/**/*.md
Options
| Flag | Description |
|---|---|
-j, --json |
Output results as JSON |
-v, --verbose |
Show individual violations and advice |
-q, --quiet |
Only print sources that fail the threshold |
-t SCORE, --threshold SCORE |
Minimum passing score (0-100). Exit 1 if any file scores below this |
-c JSONL, --config JSONL |
Path to JSONL rule configuration. Defaults to packaged settings |
-s, --score-only |
Print only numeric score output |
--counts |
Show per-rule hit counts in the summary line |
Examples
# One-line summary per file
sg draft.md
# => draft.md: 72/100 [light] (1843 words) *
# Score-only output
sg -s draft.md
# Use a custom rule config
sg -c /path/to/config.jsonl draft.md
# Verbose output with violations and advice
sg -v draft.md
# JSON for scripting
sg -j report.md | jq '.score'
# CI gate: fail if any file scores below 60
sg -t 60 docs/*.md
# Quiet mode: only show failures
sg -q -t 60 **/*.md
Exit codes
| Code | Meaning |
|---|---|
| 0 | Success (all files pass threshold, or no threshold set) |
| 1 | One or more files scored below the threshold |
| 2 | Error (bad file path, read failure, etc.) |
Fit Rule Configs (sg-fit)
Use sg-fit to fit a rule JSONL config from corpus data:
# Legacy shorthand
sg-fit TARGET_CORPUS OUTPUT
# Multi-input mode (for shell-expanded globs or many files)
sg-fit --output OUTPUT TRAIN_INPUT [TRAIN_INPUT ...]
Example:
sg-fit data.jsonl rules.fitted.jsonl
sg-fit --output rules.fitted.jsonl **/*.txt **/*.md
Optional arguments:
--init JSONL-- Start from a specific rule config JSONL instead of packaged defaults.--negative-dataset INPUT [INPUT ...]-- Add negative dataset inputs. This flag can be repeated; all negative rows are normalized to label0.--no-calibration-- Skip post-fit contrastive penalty calibration for faster fitting on large corpora.--output JSONL-- Required when you pass more than one training input.
Target corpus rows can be either:
{"text": "body of text", "label": 1}
or:
{"text": "body of text"}
If label is omitted in the target corpus, sg-fit treats it as 1 (positive/target style).
In addition to .jsonl, sg-fit accepts .txt and .md files and normalizes each file into a single training sample behind the scenes.
Installation
Requires uv.
Run without installing (recommended for MCP setups):
uvx slop-guard
# MCP server with custom rule config
uvx slop-guard -c /path/to/config.jsonl
Install persistently (gives you both slop-guard MCP server and sg CLI):
uv tool install slop-guard
Pin versions for reproducibility:
uvx slop-guard==0.3.0
Upgrade an installed tool:
uv tool upgrade slop-guard
From source
From a local checkout:
uv run slop-guard # MCP server
uv run slop-guard -c config.jsonl
uv run sg # CLI linter
uv run sg-fit data.jsonl rules.fitted.jsonl
MCP Tools
check_slop(text) -- Analyze a string. Returns JSON.
check_slop_file(file_path) -- Read a file from disk and analyze it. Same output, plus a file field.
What it catches
The linter checks for overused vocabulary (adjectives, verbs, nouns, hedging adverbs), stock phrases and filler, structural patterns (bold-header-explanation blocks, long bullet runs, triadic lists, bold-term bullet runs, bullet-heavy formatting), tone markers (meta-communication, false narrativity, sentence-opener tells, weasel phrases, AI self-disclosure), rhythm monotony (uniform sentence length), em dash and elaboration colon density, contrast pairs, setup-resolution patterns, and repeated multi-word phrases (4-8 word n-grams appearing 3+ times).
Scoring uses exponential decay: score = 100 * exp(-lambda * density), where density is the weighted penalty sum normalized per 1000 words. Claude-specific categories (contrast pairs, setup-resolution, pithy fragments) get a concentration multiplier. Repeated use of the same tic costs more than diverse violations.
Scoring bands
| Score | Band |
|---|---|
| 80-100 | Clean |
| 60-79 | Light |
| 40-59 | Moderate |
| 20-39 | Heavy |
| 0-19 | Saturated |
Output
Both tools return JSON with this structure:
score 0-100 integer
band "clean" / "light" / "moderate" / "heavy" / "saturated"
word_count integer
violations array of {type, rule, match, context, penalty}
counts per-category violation counts
total_penalty sum of all penalty values
weighted_sum after concentration multiplier
density weighted_sum per 1000 words
advice array of actionable strings, one per distinct issue
violations[].type is always "Violation" for typed records.
Benchmark snapshot
Example score distribution from benchmark/us_pd_newspapers_histogram.py on
PleIAs/US-PD-Newspapers (first 9,001 rows of one local shard):
Example score-vs-length scatter plot from
benchmark/us_pd_newspapers_scatter.py on the same shard:
Example per-rule compute-time curves from benchmark/compute-time.py +
benchmark/chart.py (annotated with the slowest rules at max length):
License
MIT
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file slop_guard-0.3.0.tar.gz.
File metadata
- Download URL: slop_guard-0.3.0.tar.gz
- Upload date:
- Size: 93.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: uv/0.10.6 {"installer":{"name":"uv","version":"0.10.6","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
7def8e2ba5488867a635576623f21c9667450792a048ecc1f538f02972f956fe
|
|
| MD5 |
f4952fee4ec937d3af1085c5524825ee
|
|
| BLAKE2b-256 |
439e57eb2f7404441d0c92e7b32228eb632d1c193c7a6189791937891ab6309a
|
File details
Details for the file slop_guard-0.3.0-py3-none-any.whl.
File metadata
- Download URL: slop_guard-0.3.0-py3-none-any.whl
- Upload date:
- Size: 69.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: uv/0.10.6 {"installer":{"name":"uv","version":"0.10.6","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
f039e1a435ba4ca9a466891240d842fff2be3651176bb56bcc281576394e2346
|
|
| MD5 |
f42be4b7de42b17f91d416691f857f51
|
|
| BLAKE2b-256 |
eb786d8b2c72c4132108e9e7ce1345026b34ba383e98ff2b323211372af10207
|