a builder's codex
codex · playbooks · Humanizer playbook

Humanizer playbook

Architectural diagram · 11 detection rule packs

Strip the AI Tells

Error severity
KillListorchestration · seamless · leverage · unlock · robust · comprehensive — delete on sight
ThroatClearing"It's worth noting that…" "Furthermore," "In conclusion," — cut entirely
NegativeParallelism"It's not X, it's Y" — the most common AI tell in the wild
ChatbotArtifacts"Certainly!" "Great question!" "I'd be happy to…" — immediate rewrite
MetaCommentarysentences that describe what's about to happen instead of happening
Warning severity
WeakVerbsutilize → use · leverage → use · facilitate → help · implement → do
AbstractNouns"a synergy of capabilities" → name the two things; combine them
ClichesAI fingerprint patterns that signal the draft wasn't written by a human with stakes
ParticiplePhrases"Having established a baseline, we then…" — construction overused in AI prose
Rewrite process
Three-layer detectionVale rule packs → semantic check → LLM rewrite
Flag errors firstfix all error-severity packs before touching warnings
Read aloud gateif a sentence sounds like it was generated, it was generated; rewrite from the reader's perspective
Per-claim factualitybinary, not Likert; each claim is verified or flagged

Audit any draft against the full anti-AI-writing canon, then rewrite the failing parts so the output reads like a sharp human typed it. Three-layer detection: rule packs (Vale) → semantic check → LLM rewrite. Single-pass operator workflow.

Source synthesis: Wikipedia WikiProject AI Cleanup field guide (rule-source-of-truth), blader/humanizer v2.5.1 (workflow + audit pattern), Cole Schaefer (Honey Copy sentence craft), Joanna Wiebe (Copyhackers conversion canon), per-claim factuality eval (Hamel + Shreya, Evals are systematic data analysis on your LLM application, start with error analysis, not tests, Build LLM-as-judge as binary true/false, one judge per pesky failure mode, and validate against human labels).

When to use

Any customer-facing draft before publish: LP, email, blog, content, ad, video script. Especially before any LLM-drafted content that the team will sign their name to.

The 11 detection rule packs

Rule packSeveritySource
KillListerroroperator-voice canon
WeakVerbswarningsentence craft
AbstractNounswarningconversion canon
ThroatClearingerroroperator-voice
ClicheswarningAI-fingerprint canon
BannedVocabularyerrorbrand voice DNA, dead AI vocabulary
NegativeParallelismerror (the big one)the most common AI tell, "It's not X, it's Y"
MetaCommentaryerror"In conclusion," "It's worth noting that…"
ParticiplePhraseswarningoverused construction in AI prose
ChatbotArtifactserror"Certainly!", "Great question!", "I'd be happy to…"
CopulaAvoidancewarningoveruse of "to be" verbs
DeadPhraseserror"delve," "unleash," "tapestry," "navigate the landscape"

Negative parallelism is THE BIG ONE because every LLM produces it dozens of times per response. Detection alone catches the gap; humanizer closes it with a rewrite layer.

Three layers

  1. Detection. Vale rule packs flag every violation with line, match, severity, and fix hint.
  2. Semantic check. LLM ingests surrounding paragraph + Vale finding + voice DNA principles → produces a single replacement sentence/clause that fixes the violation, preserves the underlying claim, matches the asset's voice. Non-destructive: writes to a sibling file for human review.
  3. Audit signature pass. After detection + optional rewrite, ask the LLM the blader/humanizer audit question:
  4. > "What makes this obviously AI generated?"

Record the answer. If it names new tells the rule packs missed, add them to the rule-pack queue. This is the auto-improvement loop.

Output

Exit codes

Composes with

Refusal patterns

Calibration

Track under taste-type "voice-enforcement" and "humanizer-rewrite." Brier signal: post-publish reader-detection rate. Look for low AI-detection rate on the published page.

Common failure modes

Why this matters

Detection alone catches the gap. Without the rewrite layer, operators get a fail report and fix by hand, that is the gate's role, not the editor's. With the humanizer, the operator gets a draft + a reviewable rewrite in one pass. The compound is read-time-to-ship reduction.

Open the interactive view →