Skip to content
Back to blog
Accuracy
April 10, 2026

AI Detection False Positives: Why Formal Writing Gets Flagged

You wrote it yourself, and the detector says it looks AI-generated. This is more common than it should be. Here is why it happens and what it actually means.

What a false positive is and why it matters

A false positive is when an AI detector classifies human-written text as AI-generated. At an individual level this is frustrating. At an institutional level (a teacher flagging a student, a publisher rejecting a freelancer, an employer questioning an applicant) a false positive can have real consequences for someone who did nothing wrong.

The false positive rate varies significantly across detector types and content categories. General-purpose detectors trained on broad internet text have false positive rates ranging from 3-15% depending on the content type being tested. For specific categories like academic writing, ESL prose, or technical documentation, that rate can climb considerably higher.

Why detectors flag human writing

AI detectors work by measuring statistical properties of text that differ between AI-generated and human-written content. The core measures:

Perplexity

How surprising each word choice is given the words before it. AI models generate low-perplexity text (predictable word choices). Human writing is higher-perplexity (more surprising).

False positive risk

Formal human writing is also low-perplexity. Academic style conventions, legal language, and technical documentation all use predictable vocabulary within their genre.

Burstiness

How much sentence length varies. Human writing has high burstiness: long sentences followed by short ones, fragments, varied structure. AI output has low burstiness: uniformly moderate sentence lengths.

False positive risk

Certain human writing styles are intentionally low-burstiness. Technical documentation, procedural writing, and legal prose use consistent sentence structures by convention.

Phrase frequency

How often certain phrases appear that are overrepresented in AI training data or AI output. Models overuse transition phrases, hedging constructions, and certain structural patterns.

False positive risk

Many of these phrases are standard in academic writing. Hedge phrases like 'it is important to note' and transition patterns like 'furthermore' and 'moreover' are taught in academic writing programs.

Vocabulary distribution

The spread of word choices. AI models favor words in the mid-frequency range, avoiding both very common and very rare words.

False positive risk

Domain-specific writing also clusters around a specialized vocabulary. A medical paper uses medical terms at high frequency; a legal brief uses legal terms. Specialized vocabulary at consistent density looks similar to AI vocabulary patterns.

Who gets flagged most often

ESL writers

Very High

Non-native English speakers often use textbook-correct sentence patterns that match AI training data closely. Formal grammar with fewer idiomatic expressions scores high on pattern detectors.

Academic writers

High

Academic writing conventions include passive voice, hedging, formal transitions, and disciplinary jargon. These overlap heavily with AI output patterns.

Technical writers

High

Step-by-step instructions, consistent terminology, and procedural structure all reduce burstiness and perplexity to AI-like levels.

Legal writers

High

Boilerplate clauses, standardized language, and formal hedging create low-perplexity, low-burstiness text that detectors misclassify.

PR and marketing writers

Medium-High

Benefit-focused, positive-framing language with consistent brand terms creates patterns similar to AI promotional content.

Journalists

Low-Medium

Inverted pyramid structure and attribution patterns can trigger some detectors, but quote variety and source specificity generally lower scores.

The academic writing problem in detail

Academic writing is the most extensively documented false positive category. A 2023 study published in PLOS One found that several commercial AI detectors misclassified 10-40% of non-native English speaker academic essays as AI-generated, compared to 1-5% for native English speakers writing on the same topics.

The specific academic writing patterns that trigger detectors:

Passive voice constructions: 'It was found that', 'The results indicate', 'This paper examines'
Hedge phrases: 'It could be argued', 'The evidence suggests', 'This may indicate'
Formal transitions: 'Furthermore', 'Moreover', 'In addition to the above', 'It is worth noting'
Disciplinary boilerplate: Standard Methods sections, APA citation patterns, abstract structure
Consistent register: Academic writing avoids informal language, slang, and tonal variation

All of these are taught as correct academic style. Students who have learned to write well academically are more likely to get flagged by AI detectors, not less.

How ensemble detection reduces false positives

Single-model detectors fail in predictable ways because they measure one or two signals. Ensemble models that run multiple independent detectors and combine their outputs have lower false positive rates because:

1

Independent failure modes cancel out

A statistical model might flag formal writing. A transformer model trained on actual AI outputs might not. When both must agree (or when scores are weighted), the statistical model's false positive does not become a verdict.

2

Coherence signals are harder to fake

AI text has distinctive coherence patterns: ideas connect too smoothly, topic transitions are unnaturally clean, the level of abstraction is too consistent. These coherence signals are not present in formal human writing even when other metrics look AI-like.

3

Metadata and structural signals add context

A deep learning model can distinguish between 'formal because trained to write formally' and 'formal because this is a research paper.' Document structure, paragraph organization, and semantic flow differ between AI and academic human writing even when surface patterns are similar.

Airno runs seven independent detectors and a fine-tuned DeBERTa-v3 model. The ensemble approach means that a single statistical signal triggering is not enough to produce a high AI score. Multiple independent detectors must agree.

How to interpret a high score on your own writing

If you run your own human-written text through an AI detector and get a high score, here is what to look at:

High pattern score, lower others

Your writing uses AI-characteristic transition phrases and hedges at above-normal density. This is a style issue, not evidence of AI generation. See which specific phrases appear in the pattern analysis and consider whether any are worth cutting.

High statistical score, others lower

Your sentence-level word choice is predictable within your genre. This is normal for technical and academic writing. The statistical model is probably miscalibrated for your content type.

All detectors elevated (70%+)

This is less likely to be a pure false positive. Review the text carefully. If you used AI for any part of the drafting process, that may be reflected here.

Score varies by section

Check if different sections score differently. If one section is notably higher, that section may have AI influence even if the rest is human-written.

What detectors cannot do

No AI detector can be certain. Detection scores are probabilistic, not forensic. The same text can score differently across detectors, on the same detector at different times, or when minor variations are made to the input. Any institutional policy that treats a single detection score as definitive evidence is misapplying the technology.

The appropriate use of AI detection is as one input into a broader review. A high score is a reason to look more carefully, not a verdict. A low score is evidence of lower AI likelihood, not proof of human authorship.

For more context on detection accuracy and how different detectors compare, see Why AI Detectors Fail on Paraphrased Text and Airno vs GPTZero.

See which detectors actually fired

Airno shows a breakdown by detector type so you can see exactly what is driving the score. Free, no account needed.

Try Airno free