AI Writing Detection Reference
Expert-level knowledge base for detecting AI-generated text, compiled from academic research, commercial detection tools, and empirical analysis.
Quick Reference: High-Confidence Signals
These indicators strongly suggest AI authorship when found together:
Vocabulary Red Flags
High-signal words (50-700x more common in AI text):
-
"delve", "tapestry", "nuanced", "multifaceted", "underscore"
-
"intricate interplay", "played a crucial role", "complex and multifaceted"
-
"paramount", "pivotal", "meticulous", "holistic", "robust"
-
"stands/serves as", "marking a pivotal moment", "underscores its importance"
Overused phrases:
-
"It's important to note that..."
-
"In today's fast-paced world..."
-
"At its core..."
-
"Without further ado..."
-
"Let me explain..."
See reference/vocabulary-patterns.md for complete lists.
Structural Red Flags
-
Uniform sentence lengths: 12-18 words consistently (low burstiness)
-
Tricolon structures: "research, collaboration, and problem-solving"
-
Em dash overuse: AI uses em dashes in a formulaic way to mimic "punched up" sales writing, especially in parallelisms ("it's not X — it's Y"); swapping punctuation doesn't fix the underlying emphasis pattern
-
Perfect paragraph uniformity: All paragraphs same approximate length
-
Template conclusions: "In summary...", "In conclusion..."
-
Negative parallelisms: "It's not about X; it's about Y"
-
Elegant variation: Cycling through synonyms to avoid repetition
-
False ranges: "From X to Y" with incoherent endpoints
See reference/structural-patterns.md for details.
Content Red Flags
-
Importance puffery: "marking a pivotal moment in history"
-
Ecosystem/conservation claims without citations
-
"Challenges and Future" sections following rigid formula
-
Promotional language: "nestled in", "stunning natural beauty", "boasts"
-
Superficial analyses: "-ing" phrases attributing significance to facts
See reference/content-patterns.md for details.
Formatting Red Flags
-
Title Case in all section headings
-
Excessive boldface (every key term bolded)
-
Inline-header lists: Bold Header: description pattern
-
Emojis in formal content or headings
-
Subject lines in non-email contexts
See reference/formatting-patterns.md for details.
Markup Red Flags (Definitive)
-
turn0search0, turn0image0: ChatGPT reference markers
-
contentReference[oaicite:]: ChatGPT reference bugs
-
utm_source=chatgpt.com: URL tracking (definitive)
-
Markdown in wikitext: ## headers, bold, text
-
grok_card XML tags: Grok/X specific
See reference/markup-artifacts.md for details.
Citation Red Flags
-
Broken external links that never existed (no archive)
-
Invalid DOIs/ISBNs: Checksum failures
-
Declared but unused references: Cite errors
-
Placeholder values: url=URL , date=2025-XX-XX
See reference/citation-patterns.md for details.
Tone Red Flags
-
Passive and detached voice throughout
-
Absence of first-person pronouns where expected
-
Consistent formality with no stylistic variation
-
Over-politeness and excessive hedging
Detection Methodology
Multi-Layer Analysis Approach
Layer 1: Technical Artifact Scan (Definitive)
-
Check for turn0search/oaicite markers (ChatGPT)
-
Check for utm_source=chatgpt.com in URLs
-
Check for grok_card tags (Grok)
-
Check for Markdown in non-Markdown contexts
-
If found: Definitive AI involvement
Layer 2: Vocabulary Pattern Matching
-
Scan for overused AI words/phrases
-
Count frequency of flagged terms
-
Look for clusters of high-signal vocabulary
-
Check for importance/symbolism phrases
Layer 3: Structural Analysis
-
Observe sentence length variation (uniform = AI signal)
-
Check paragraph uniformity
-
Identify repetitive syntactic templates (tricolons, negative parallelisms)
-
Look for elegant variation (synonym cycling)
-
Check for false ranges
Layer 4: Content Pattern Analysis
-
Check for importance puffery and promotional language
-
Look for "Challenges and Future" formula
-
Check for ecosystem/conservation claims without citations
-
Identify superficial analyses with "-ing" attributions
Layer 5: Citation Verification
-
Test external links - do they exist?
-
Verify DOI/ISBN checksums
-
Check for declared but unused references
-
Look for placeholder values
Layer 6: Formatting Analysis
-
Check heading capitalization (Title Case = signal)
-
Count bold phrases per paragraph
-
Look for inline-header list patterns
-
Check for emojis in formal content
Layer 7: Stylometric Observation
-
Pronoun usage patterns (missing first-person?)
-
Tone consistency (too uniform = AI signal)
-
Punctuation patterns (em dash overuse? curly quotes?)
Layer 8: Coherence Check
-
Do paragraphs build a coherent argument?
-
Are concepts repeated with different words?
-
Do transitions actually connect ideas?
Layer 9: Confidence Scoring
-
Weight multiple signals together
-
Require corroborating evidence (3+ signals minimum)
-
Apply context-specific adjustments
-
Check for mitigating factors (human signals)
-
Consider ineffective indicators (don't use them)
Model-Specific Patterns
Different AI models have distinct "fingerprints":
Model Key Tells Technical Artifacts
ChatGPT/GPT-4 "delve" (pre-2025), "tapestry", tricolons, em dashes, curly quotes turn0search, oaicite, utm_source=chatgpt.com
Claude Analytical structure, extended analogies, cautious qualifications None (uses straight quotes, no tracking)
Gemini Conversational synthesis, fact-dense paragraphs None (uses straight quotes, no tracking)
DeepSeek Similar to ChatGPT, curly quotes Curly quotation marks
Grok X/Twitter integration <grok_card> XML tags
Perplexity Source-focused output [attached_file:1] , [web:1] tags
Important dates:
-
ChatGPT launched: November 30, 2022 (text before this is almost certainly human)
-
"delve" usage dropped: 2025 (still signals pre-2025 ChatGPT)
See reference/model-fingerprints.md for detailed model patterns.
False Positive Prevention
Critical requirements:
-
Minimum 200 words for reliable analysis
-
Never flag on single indicators alone
-
Use ensemble scoring (multiple signals required)
High false-positive risk groups:
-
Non-native English speakers (61% false positive rate in research)
-
Technical/formal writing
-
Neurodivergent writers
-
Content using grammar correction tools
Ineffective indicators (do NOT rely on these):
-
Perfect grammar alone
-
"Bland" or "robotic" prose
-
"Fancy" or unusual vocabulary
-
Letter-like formatting alone
-
Conjunctions starting sentences
Signs of human writing:
-
Text from before November 30, 2022
-
Ability to explain editorial choices
-
Personal anecdotes with verifiable details
-
Minor errors and natural quirks
See reference/false-positive-prevention.md for detailed guidance.
Analysis Output Format
Structure findings as:
Overall Assessment: [Likely AI / Possibly AI / Likely Human / Inconclusive] Confidence: [Low / Medium / High]
Summary: 2-3 sentence overview
Evidence Found:
- [Category]: [Specific indicator] - "[Quote from text]"
- [Category]: [Specific indicator] - "[Quote from text]"
Mitigating Factors: [Elements suggesting human authorship]
Caveats: [Limitations, alternative explanations]
Key Principles
-
No certainty claims - AI detection is probabilistic
-
Multiple signals required - Single indicators prove nothing
-
Context matters - Academic writing differs from blogs
-
Stakes awareness - False accusations cause real harm
-
Evolving field - Detection methods require constant updates
Reference Files
-
vocabulary-patterns.md - Complete word/phrase lists with frequencies
-
structural-patterns.md - Sentence, paragraph, and discourse patterns
-
content-patterns.md - Importance puffery, promotional language, content tells
-
formatting-patterns.md - Title case, boldface, emojis, visual patterns
-
markup-artifacts.md - Technical artifacts: turn0search, oaicite, Markdown, tracking
-
citation-patterns.md - Broken links, invalid identifiers, hallucinated references
-
model-fingerprints.md - GPT, Claude, Gemini, Grok, Perplexity specific tells
-
false-positive-prevention.md - Avoiding false accusations, ineffective indicators
Sources
This knowledge base synthesizes research from:
-
Stanford HAI (DetectGPT, bias studies)
-
GPTZero, Originality.ai, Turnitin, Pangram methodologies
-
Academic papers on stylometry and discourse analysis
-
Empirical studies on detection accuracy and limitations
-
Wikipedia:WikiProject AI Cleanup field guide (2025)
-
Community-documented patterns from Wikipedia editing