Fact Checker Investigator
Overview
Systematically verify claims, validate sources, cross-reference data, and investigate discrepancies. Ensure research integrity by distinguishing verified facts from assumptions, interpretations, and fabrications.
Core Mission
Never publish unverified claims as facts.
Protect research integrity by:
- Verifying all quotes have real sources
- Confirming data and metrics are traceable
- Identifying assumptions masquerading as facts
- Cross-referencing conflicting information
- Flagging fabricated or questionable content
Verification Checklist
Stakeholder quotes
Required for every quote:
- Source file exists and is accessible
- Line number is approximately correct (±5 lines acceptable)
- Quote matches source text (exact or accurate paraphrase)
- Speaker attribution is correct (name, role)
- Context preserved (not taken out of context)
Red flags:
- Generic quotes without specific sources ("Artists say...")
- Convenient quotes that perfectly support argument
- Attribution to roles without names ("A story artist mentioned...")
- Line numbers that don't exist in cited file
Data and metrics
Required for every number:
- Source document cited
- Measurement definition clear (what's being counted, timeframe)
- Calculation method traceable
- Context provided (sample size, conditions)
Red flags:
- Round numbers without source (40%, "most", "many")
- Precise statistics without methodology
- Comparisons without baseline
- Trend claims without time series data
Meeting decisions
Required for every decision:
- What was decided (specific, actionable)
- Who made decision (decision owner by name)
- When decided (meeting date)
- Source documented (meeting notes with line number)
- Rationale captured (why this choice)
Red flags:
- Passive voice decisions ("It was decided...")
- Vague outcomes ("Team agreed to explore...")
- Decisions without decision-maker names
- No source documentation
Investigation Workflow
1. Document audit
Read target document and flag all claims:
- Direct quotes
- Data/metrics
- Decisions
- Timelines
- Stakeholder preferences
- Technical specifications
2. Source verification
For each flagged claim:
**Claim:** [The assertion being made]
**Source cited:** [What document/line is referenced]
**Verification:**
- [ ] Source file exists
- [ ] Content matches claim
- [ ] Context appropriate
**Status:** ✅ Verified | ⚠️ Questionable | ❌ Unverified
**Notes:** [Any discrepancies or concerns]
3. Cross-reference investigation
When conflicting information exists:
- Identify all sources making competing claims
- Check dates (is one source more recent/authoritative?)
- Examine context (are they talking about same thing?)
- Note uncertainty ("Sources conflict: A says X, B says Y")
4. Gap identification
Flag unverified claims:
- Missing source: Claim has no citation
- Inaccessible source: Citation points to non-existent file
- Misattribution: Quote attributed to wrong person
- Out of context: Quote meaning distorted
- Fabrication suspected: Claim seems invented
5. Recommendation
For each issue found:
- Fix: If correction is clear (wrong line number, typo)
- Verify: If source exists but needs checking
- Remove: If claim is unverifiable and non-critical
- Clarify: If claim needs "assumption" or "hypothesis" qualifier
Investigation Script
Use scripts/verify_citations.py to automate source checking:
python scripts/verify_citations.py document.md --check-quotes --check-line-numbers
Output:
- List of all quotes with verification status
- Broken citations (non-existent files/lines)
- Suspicious patterns (many quotes from same line, generic attributions)
Common Scenarios
Scenario 1: Pre-publication review
Context: Research report ready to share with stakeholders
Process:
- Run automated citation checker
- Manually verify high-impact claims (decisions, recommendations, key data)
- Flag assumptions that read like facts
- Request clarification for questionable content
- Sign off only when critical claims verified
Scenario 2: Conflicting information
Context: Two documents make contradictory claims
Process:
- Identify exact nature of conflict
- Check source recency (newer may supersede)
- Check source authority (who's closer to truth?)
- Document both perspectives if unresolvable
- Note explicitly: "Sources conflict on this point"
Scenario 3: Suspicious document
Context: Document has markers of fabricated content
Markers:
- Too many "perfect" quotes
- Convenient data supporting conclusions
- Generic attributions
- No line numbers or vague citations
Process:
- Flag all suspicious claims
- Request source verification from author
- Do not approve until verification provided
- Consider full document review if multiple issues
Output: Verification Report
# Fact-Check Report: [Document Name]
**Date:** YYYY-MM-DD
**Reviewer:** [Your name]
**Status:** ✅ Approved | ⚠️ Approved with notes | ❌ Revision needed
## Summary
[Overall assessment: strong/weak evidence, major issues found]
## Critical Issues (Must Fix)
1. **[Claim]** - [Issue: fabricated/misattributed/unverified]
- Location: [File:line]
- Recommendation: [Fix/Remove/Verify]
## Minor Issues (Should Fix)
[List of non-critical but important corrections]
## Verified Claims (High Confidence)
[Key claims that passed rigorous checking - builds trust]
## Assumptions Flagged
[Claims that should be marked as assumptions/hypotheses, not facts]
Resources
scripts/
verify_citations.py- Automated citation checking tool
references/
verification-standards.md- Detailed criteria for different claim typesinvestigation-techniques.md- Advanced cross-referencing methods
assets/
verification-report-template.md- Standard report format
Red Line Principles
Never compromise on:
- Stakeholder quotes must have real, traceable sources
- Data must cite methodology and source
- Decisions must name decision-makers
- Assumptions must be labeled as such
- Conflicting evidence must be acknowledged
When in doubt: Mark as "unverified" and request clarification. Better to delay publication than publish fiction.