How to check AI writing: a practical workflow
The goal is signal clarity, not punishment theater. If you need to check whether a piece of writing looks AI-generated, this workflow gives you a fair process that combines diagnostic evidence with context — so the result is useful whether the answer is "yes," "no," or "it's complicated."
Quick answer
Five steps, in order. Do not skip to step five.
- Collect enough text — at least 150 words for a stable read.
- Look for clusters of AI writing patterns, not one magic giveaway.
- Run a transparent tool like the WROITER Diagnostic so you can see what triggered the score.
- Compare the result against genre, revision history, and context.
- Decide: revise, investigate, or move on. Not: accuse.
Step 1: collect enough text
The diagnostic accepts 50 words minimum. But 50 words is a single paragraph — barely enough for rhythm patterns to emerge and short enough for one stock phrase to dominate the score. At 150 words and above, the tool has room to see whether a pattern repeats or was a one-off quirk.
If you are checking something short — a product description, an email, a social post — keep in mind that the result is less stable. A score of 40 on a 60-word sample means less than a score of 40 on 400 words.
Step 2: inspect the pattern evidence first
Before you react to the number, read what triggered it. The diagnostic flags specific pattern families: stock phrases, throat-clearing intros, flat rhythm, over-signposting, rhetorical pivot crutches. Each flag links to a pattern you can find and verify in the actual text.
This is the step that changes the question from "is it AI?" to "what am I actually seeing?" A text might score 55 because it has three stock phrases and mild rhythm compression. Or it might score 55 because the rhythm is extremely flat across the whole sample. Those are different situations that call for different responses.
For full definitions and examples of every pattern, see the Pattern Library. For the visual-recognition version (what to notice when reading without a tool), see How to Spot AI Writing.
Step 3: run a transparent diagnostic
Most AI writing checkers give you a percentage and nothing else. That is not useful. It is a confidence trick dressed up as a confidence score.
The WROITER Diagnostic returns the score plus everything behind it: rhythm metrics (average sentence length, standard deviation, burstiness), flag-level detail for every triggered pattern, and notes explaining what each flag means. You can verify every flag in the text yourself. If you cannot find what the tool is reacting to, the score is not actionable.
Step 4: compare the result against context
Context matters as much as the score — sometimes more. Questions worth asking:
- Does the writer's genre naturally compress variation? Academic abstracts, legal prose, and product copy are formal by design. They will score higher than personal essays even when fully human-written.
- Is this text heavily edited? Multiple editing passes smooth out quirks. That smoothness looks like the same flatness detectors flag in AI output.
- Is the writer working in a second language? Non-native writers often choose safe, common phrasings — exactly the kind of vocabulary and structure that triggers pattern flags.
- Is there a revision history? Drafts, outlines, and version notes are stronger evidence than any detector score.
The limitations page documents the genres and situations where false positives are most common. Read it before you treat a score as meaningful in any process with consequences.
Step 5: decide on the next action
You have the score, the flags, and the context. Now what?
Usually one of three things:
- Revise. The patterns are real and the text would be better without them. This is the most common outcome and the one that helps the writer most.
- Investigate. The score is high, the patterns are dense, and the context does not explain them. Ask for drafts, revision history, or process notes. A conversation, not an accusation.
- Move on. The evidence is thin, the context explains the patterns, or the score is low. Not every check needs to find something wrong.
The one thing you should not do: jump from the score to a conclusion without completing steps 2 through 4. That is how false positives become false accusations.
Common mistakes
- Too little text. Checking 50 words and treating the result as definitive. Short samples are unstable.
- Skipping the flags. Looking at the number, ignoring the pattern evidence. The number is a summary. The flags are the substance.
- Confusing polish with AI. Clean grammar, formal tone, and precise vocabulary are not signs of AI writing. They are signs of careful writing. If that distinction is tripping you up, read the False Positive Hall of Fame.
- Ignoring context. A score without genre, revision, and policy context is a number floating in air. It means nothing until you attach it to something.
Interpretation
A higher score means stronger overlap with AI-typical surface patterns. It does not prove authorship, cheating, or intent. For score ranges and the full interpretation boundary, see How It Works. For the broader reliability case: Do AI Detectors Work? and the False Positive Hall of Fame.
Try it
Paste your text, read the flags, check the context, then decide. That is the whole workflow.