Skip to main content

Rule-based prose linter for formulaic AI writing patterns.

Project description

slop-guard

A rule-based prose linter that scores text 0--100 for formulaic AI writing patterns. No LLM judge, no API calls. Purely programmatic.

It runs ~80 compiled patterns against your text and returns a numeric score, a list of specific violations with surrounding context, and concrete advice for each hit.

Add to Your Agent

Claude Code

Add from the command line:

claude mcp add slop-guard -- uvx slop-guard
# Optional custom rule config:
claude mcp add slop-guard -- uvx slop-guard -c /path/to/config.jsonl

Add to your .mcp.json:

{
  "mcpServers": {
    "slop-guard": {
      "command": "uvx",
      "args": ["slop-guard"]
    }
  }
}

With a custom rule config:

{
  "mcpServers": {
    "slop-guard": {
      "command": "uvx",
      "args": ["slop-guard", "-c", "/path/to/config.jsonl"]
    }
  }
}

Codex

Add from the command line:

codex mcp add slop-guard -- uvx slop-guard
# Optional custom rule config:
codex mcp add slop-guard -- uvx slop-guard -c /path/to/config.jsonl

Add to your ~/.codex/config.toml:

[mcp_servers.slop-guard]
command = "uvx"
args = ["slop-guard"]

With a custom rule config:

[mcp_servers.slop-guard]
command = "uvx"
args = ["slop-guard", "-c", "/path/to/config.jsonl"]

If you want a fixed release, pin it in args, for example: ["slop-guard==0.3.0"].

CLI

The sg command lints prose files from the terminal. No API keys, no network calls.

Quick start

# Run without installing
uvx --from slop-guard sg README.md

# Or install it
uv tool install slop-guard
sg README.md

Usage

sg [OPTIONS] INPUT [INPUT ...]

sg requires at least one input. Each input can be a file path, - for stdin, or quoted inline prose text:

sg "This is some test text"
echo "This is a crucial paradigm shift." | sg -

Lint multiple files at once (shell-level glob expansion):

sg docs/*.md README.md
sg path/**/*.md

Options

Flag Description
-j, --json Output results as JSON
-v, --verbose Show individual violations and advice
-q, --quiet Only print sources that fail the threshold
-t SCORE, --threshold SCORE Minimum passing score (0-100). Exit 1 if any file scores below this
-c JSONL, --config JSONL Path to JSONL rule configuration. Defaults to packaged settings
-s, --score-only Print only numeric score output
--counts Show per-rule hit counts in the summary line

Examples

# One-line summary per file
sg draft.md
# => draft.md: 72/100 [light] (1843 words) *

# Score-only output
sg -s draft.md

# Use a custom rule config
sg -c /path/to/config.jsonl draft.md

# Verbose output with violations and advice
sg -v draft.md

# JSON for scripting
sg -j report.md | jq '.score'

# CI gate: fail if any file scores below 60
sg -t 60 docs/*.md

# Quiet mode: only show failures
sg -q -t 60 **/*.md

Exit codes

Code Meaning
0 Success (all files pass threshold, or no threshold set)
1 One or more files scored below the threshold
2 Error (bad file path, read failure, etc.)

Fit Rule Configs (sg-fit)

Use sg-fit to fit a rule JSONL config from corpus data:

# Legacy shorthand
sg-fit TARGET_CORPUS OUTPUT

# Multi-input mode (for shell-expanded globs or many files)
sg-fit --output OUTPUT TRAIN_INPUT [TRAIN_INPUT ...]

Example:

sg-fit data.jsonl rules.fitted.jsonl
sg-fit --output rules.fitted.jsonl **/*.txt **/*.md

Optional arguments:

  • --init JSONL -- Start from a specific rule config JSONL instead of packaged defaults.
  • --negative-dataset INPUT [INPUT ...] -- Add negative dataset inputs. This flag can be repeated; all negative rows are normalized to label 0.
  • --no-calibration -- Skip post-fit contrastive penalty calibration for faster fitting on large corpora.
  • --output JSONL -- Required when you pass more than one training input.

Target corpus rows can be either:

{"text": "body of text", "label": 1}

or:

{"text": "body of text"}

If label is omitted in the target corpus, sg-fit treats it as 1 (positive/target style).

In addition to .jsonl, sg-fit accepts .txt and .md files and normalizes each file into a single training sample behind the scenes.

Installation

Requires uv.

Run without installing (recommended for MCP setups):

uvx slop-guard
# MCP server with custom rule config
uvx slop-guard -c /path/to/config.jsonl

Install persistently (gives you both slop-guard MCP server and sg CLI):

uv tool install slop-guard

Pin versions for reproducibility:

uvx slop-guard==0.3.0

Upgrade an installed tool:

uv tool upgrade slop-guard

From source

From a local checkout:

uv run slop-guard               # MCP server
uv run slop-guard -c config.jsonl
uv run sg            # CLI linter
uv run sg-fit data.jsonl rules.fitted.jsonl

MCP Tools

check_slop(text) -- Analyze a string. Returns JSON.

check_slop_file(file_path) -- Read a file from disk and analyze it. Same output, plus a file field.

What it catches

The linter checks for overused vocabulary (adjectives, verbs, nouns, hedging adverbs), stock phrases and filler, structural patterns (bold-header-explanation blocks, long bullet runs, triadic lists, bold-term bullet runs, bullet-heavy formatting), tone markers (meta-communication, false narrativity, sentence-opener tells, weasel phrases, AI self-disclosure), rhythm monotony (uniform sentence length), em dash and elaboration colon density, contrast pairs, setup-resolution patterns, and repeated multi-word phrases (4-8 word n-grams appearing 3+ times).

Scoring uses exponential decay: score = 100 * exp(-lambda * density), where density is the weighted penalty sum normalized per 1000 words. Claude-specific categories (contrast pairs, setup-resolution, pithy fragments) get a concentration multiplier. Repeated use of the same tic costs more than diverse violations.

Scoring bands

Score Band
80-100 Clean
60-79 Light
40-59 Moderate
20-39 Heavy
0-19 Saturated

Output

Both tools return JSON with this structure:

score          0-100 integer
band           "clean" / "light" / "moderate" / "heavy" / "saturated"
word_count     integer
violations     array of {type, rule, match, context, penalty}
counts         per-category violation counts
total_penalty  sum of all penalty values
weighted_sum   after concentration multiplier
density        weighted_sum per 1000 words
advice         array of actionable strings, one per distinct issue

violations[].type is always "Violation" for typed records.

Benchmark snapshot

Example score distribution from benchmark/us_pd_newspapers_histogram.py on PleIAs/US-PD-Newspapers (first 9,001 rows of one local shard):

slop-guard score histogram

Example score-vs-length scatter plot from benchmark/us_pd_newspapers_scatter.py on the same shard:

slop-guard score vs length scatter

Example per-rule compute-time curves from benchmark/compute-time.py + benchmark/chart.py (annotated with the slowest rules at max length):

slop-guard per-rule compute time

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

slop_guard-0.3.0.tar.gz (93.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

slop_guard-0.3.0-py3-none-any.whl (69.3 kB view details)

Uploaded Python 3

File details

Details for the file slop_guard-0.3.0.tar.gz.

File metadata

  • Download URL: slop_guard-0.3.0.tar.gz
  • Upload date:
  • Size: 93.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: uv/0.10.6 {"installer":{"name":"uv","version":"0.10.6","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for slop_guard-0.3.0.tar.gz
Algorithm Hash digest
SHA256 7def8e2ba5488867a635576623f21c9667450792a048ecc1f538f02972f956fe
MD5 f4952fee4ec937d3af1085c5524825ee
BLAKE2b-256 439e57eb2f7404441d0c92e7b32228eb632d1c193c7a6189791937891ab6309a

See more details on using hashes here.

File details

Details for the file slop_guard-0.3.0-py3-none-any.whl.

File metadata

  • Download URL: slop_guard-0.3.0-py3-none-any.whl
  • Upload date:
  • Size: 69.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: uv/0.10.6 {"installer":{"name":"uv","version":"0.10.6","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for slop_guard-0.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 f039e1a435ba4ca9a466891240d842fff2be3651176bb56bcc281576394e2346
MD5 f42be4b7de42b17f91d416691f857f51
BLAKE2b-256 eb786d8b2c72c4132108e9e7ce1345026b34ba383e98ff2b323211372af10207

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page