Tutorial
How AI Detectors Work: The Science Behind AI Detection Tools
A practical explanation of how AI text detectors work, what they measure, why they make mistakes, and what this means for writers who use AI assistance.
The two core measurements: perplexity and burstiness
Most AI text detectors measure two core statistical properties of text: perplexity and burstiness. Understanding what these mean is essential for understanding both why detectors work and why they make mistakes.
Perplexity measures how predictable text is. A language model assigns a probability to each word given the words before it. High-perplexity text is surprising — the words chosen are less probable, more unexpected. Low-perplexity text is predictable — the words are exactly what a model would predict. AI-generated text, by definition, consists of high-probability word choices, so it tends to have lower perplexity than human writing.
Burstiness measures how much sentence length varies throughout a text. Human writing naturally bursts between very short and very long sentences. AI writing tends toward uniform sentence length, producing low burstiness. This is one of the most reliable signals because it is hard to fake systematically.
How detectors use these measurements
A detector calculates perplexity and burstiness scores for a submitted text and compares them to distributions established during training — typically using large corpora of known human-written and AI-generated texts. The resulting score is a probability estimate: this text has properties consistent with X% of AI-generated texts in our training data.
More sophisticated detectors also look for specific AI-signature phrases ("It is important to note that," "In conclusion," "Furthermore"), structural patterns like perfect paragraph closure, and domain-specific variations in how AI models handle different topics.
Some detectors additionally measure text entropy and use watermarking where possible. OpenAI and some other model providers have explored embedding statistical watermarks in their output — subtle biases in token selection that can be detected by tools with access to the watermarking key. This approach is more robust than pattern-matching but requires participation from the model provider.
Why AI detectors make significant errors
AI detectors have meaningful false positive rates (human text flagged as AI) and false negative rates (AI text that passes as human). Both types of errors are common enough to make detector scores unreliable as definitive evidence.
False positives are caused by: non-native English speakers who write more predictably; technical writing and documentation which use formal, predictable structures; highly polished human writing that uses consistent style; and writing in genres with conventional forms (academic, legal, medical).
False negatives are caused by: AI text that was heavily edited by humans; prompting strategies that deliberately introduce variation; AI models trained specifically to produce more human-like output; and domains where AI and human writing are naturally similar.
- False positive rate: typically 5–15% depending on the detector and text type
- False negative rate: highly variable — heavily edited AI text often passes completely
- Academic impact: Turnitin, GPTZero, and similar tools have acknowledged meaningful error rates
- Legal status: No court or institution treats AI detection as definitive proof
What this means for writers who use AI assistance
The practical implications are clear. First, never rely on a single detector result as definitive evidence of anything — neither that content is AI-generated nor that it is human-written. Second, if you receive a false positive on your own human-written work, you now know what to say: detector scores are probabilistic estimates with known error rates, not proof.
Third, and most importantly: focus on quality rather than score optimization. The patterns that make AI text detectable — uniform structure, low specificity, generic phrasing — are also the patterns that make it less useful and less engaging for readers. Fixing those patterns to produce genuinely better writing naturally reduces AI detection scores as a byproduct.
Tools like the AI Detector Remover and AI Humanizer improve content quality by targeting these exact patterns. The result is both more natural writing and typically lower AI detection scores — because the underlying quality improvement is what drives both outcomes.
The specific detectors and their approaches
Turnitin's AI detection uses a combination of perplexity, burstiness, and pattern matching, with the advantage of a massive training corpus from academic submissions. It is integrated directly into academic workflows, which makes it the most consequential detector for students. False positive rate is reportedly around 4%, but the company acknowledges individual submissions can be misclassified.
GPTZero uses perplexity and burstiness as its primary signals, with an additional document-level consistency check. It is commonly used by educators and publishers. It has several detector options at different sensitivity thresholds.
Originality.ai targets professional publishers and SEO teams. It combines AI detection with plagiarism checking and uses a higher threshold for flagging to reduce false positives in professional content workflows.
FAQ
No. AI detectors are probabilistic tools with meaningful error rates. They can be useful signals in a broader review process, but they are not reliable enough to serve as definitive evidence that content is or is not AI-generated.
Yes, but the most reliable way is genuine quality improvement — making the writing more specific, varied, and human in quality — rather than technical evasion. Detection tools continuously update their models as evasion patterns become known.
No. Running your own writing through a detector simply produces a probabilistic assessment based on its current training data. The assessment tells you something about the writing patterns present, which can be useful editorial feedback.
Try the related tool
Check pasted text for AI-like writing patterns and use the result as an editorial review signal before rewriting, editing, or publishing.
Open AI Text DetectorSupporting pages
Related articles
GPTZero, Turnitin, Originality.ai, and other AI detectors compared by accuracy, use case, and false positive rate. Which detector should you use and when?
Read articleTurnitin AI detection uses perplexity and writing patterns to flag AI-generated content. Here is how it works, where it makes mistakes, and what to do if you receive a high score.
Read articleAI writing sounds robotic because of identifiable patterns: uniform sentence length, predictable transitions, low specificity, and generic phrasing. Here is what each pattern looks like and how to fix it.
Read articleA quality-focused guide to improving AI-assisted drafts without detector-bypass claims or shallow paraphrasing.
Read article