Principles of Evaluation
AI text detectors do not prove authorship. They estimate likelihood based on patterns that can fail, especially on short text, multilingual writing, heavily edited content, or intentionally humanized AI output.
This methodology is built around two realities:
- False positives cause the most harm.
If a detector wrongly flags a human author, the consequences can be serious (academic penalties, reputational damage, contract disputes). So we design tests and scoring to aggressively penalize false positives. - Robustness matters more than perfect lab accuracy.
In the real world, AI text is rarely raw. It is edited, paraphrased, rewritten, mixed with human text, translated, or formatted. Our benchmarks include these conditions because they are the normal operating environment.
Core principles
- Risk scoring, not verdicts. We treat detector output as a signal in a broader decision process (draft history, citations, interviews/oral defenses, provenance data).
- Prefer conservative classification. When in doubt, we avoid AI-generated labels unless confidence is supported by consistent tool behavior and sufficient text length.
- Reproducibility. Same dataset version, same preprocessing rules, same reporting format across tools.
- Fairness and robustness. We actively test where detectors are known to fail (non-native English, paraphrase/humanization attacks, style shifts).
Criteria and Weights
We score tools on a 100-point scale using weighted criteria. The weights reflect practical risk management: avoiding false accusations and handling real-world edits.
1) Reliability (60 points total)
A. False positives on human text (30 points)
- We measure the false positive rate (FPR) on verified human samples.
- We also track confident false positives (high-confidence AI labels on human text) because these are the most dangerous failures.
B. Detection on pure AI text (15 points)
- We measure true positive rate (TPR) for clean, unedited AI outputs.
- This matters, but less than false positives, because most harm comes from mislabeling humans.
C. Mixed authorship sensitivity (15 points)
- We test mixed documents (for example, 20 to 50 percent AI-inserted sections) because many real documents are hybrid.
- Tools should avoid overconfident fully AI labels when only small sections are machine-generated.
2) Robustness (25 points total)
D. Paraphrase / humanization resistance (15 points)
- We test both simple paraphrasing and stronger humanization edits.
- Robust tools degrade gracefully rather than collapsing into random outputs.
E. Domain and style generalization (10 points)
- We include multiple genres (blog posts, formal essays, product copy, emails).
- Tools should generalize across domains without extreme swings.
3) Product usefulness (15 points total)
F. Reporting and explainability (6 points)
- Does it provide segment-level analysis, transparency, and exportable reports?
G. Workflow and integrations (5 points)
- Batch testing, API access, LMS/WP integrations, team features.
H. Pricing clarity (4 points)
- Predictable pricing, clear quotas/credits, no hidden constraints.
Important note on accuracy claims: Vendor-reported accuracy is recorded but not scored unless we can reproduce performance on our dataset. Vendor claims are labeled as vendor claims, not as independent results.
Testing Process
This is the protocol we follow for each tool review and comparison.
Step 1 – Define the scope (so results are honest)
- Language: English-first. Many detectors degrade in non-English text; we do not claim multilingual reliability unless separately tested.
- Text length: We avoid drawing conclusions from very short inputs. Many classifiers are unreliable on short text, so our core dataset uses longer samples.
Step 2 – Build a versioned dataset (ground truth matters)
We maintain a versioned dataset (for example, v1.0, v1.1) so results remain comparable over time.
Minimum dataset composition (baseline for reviews)
- Human-written (verified): 10 samples. Sources: consenting authors, internal writing, or public-domain texts where attribution is clear.
- AI-generated (pure): 10 samples generated across at least two different model families and multiple prompts.
- AI plus edits (humanized/paraphrased): 10 samples, including light paraphrase, heavy rewrite, and style-shift edits.
Optional expansion samples (used for higher-confidence reviews)
- Mixed authorship documents (human + AI sections at known ratios)
- Non-native English samples (separate fairness check)
- Topic splits (so training-topic leakage is less likely)
Step 3 – Normalize inputs (consistent preprocessing)
- Remove tracking metadata and headers/footers unrelated to content.
- Keep quotes and citations (then optionally run a second pass without them to test sensitivity).
- Preserve paragraph boundaries and punctuation.
We document any preprocessing. If we touch the text, we say so.
Step 4 – Run tools under consistent conditions
For each detector, we use default settings unless the tool provides a clear, documented recommended mode.
We record:
- raw score(s)
- tool label (for example, likely AI)
- confidence band (if provided)
- segment-level highlights
- time to result and input limits
We run tests in a short time window (same day if possible) to minimize model/version drift.
Step 5 – Compute metrics that reflect real risk
We report:
- False Positive Rate (FPR): percent of human samples labeled AI
- True Positive Rate (TPR / recall): percent of AI samples labeled AI
- Precision: when the tool says AI, how often it is correct
- Confident error rate: high-confidence wrong labels (especially on humans)
Because many real contexts have low prevalence of AI cheating, we include a base-rate sanity check in reports. Even a low FPR can produce many false accusations when AI prevalence is low.
Step 6 – Robustness tests (the real-world edit gauntlet)
We evaluate how performance changes under:
- Paraphrase: same meaning, different surface form
- Humanization: deliberate editing to bypass detectors
- Style shift: formal to casual, or vice versa
- Hybrid documents: known percentage of AI sections inside human text
A tool that only works on raw AI text but fails under light editing gets penalized heavily.
Step 7 – Publish results with constraints and reproducibility
Every review includes:
- dataset size and version
- what was tested (and what was not)
- limitations (language, length, domains)
- conflicts of interest (affiliate links, sponsorships)
Rules for Updating Results
We update methodology and tool results on a predictable schedule.
Scheduled updates
- Quarterly updates are the default (methodology + tool retests if needed).
- Each update produces a methodology version number, a dataset version number, and a short changelog describing what changed and why.
Trigger-based updates (out of cycle)
We update sooner if:
- a detector releases a major model change (new engine or scoring system)
- a major LLM release changes the landscape
- we discover a systematic error in testing or labeling
- a tool materially changes pricing, limits, or core features
Backward compatibility
We keep older results visible when possible, labeled clearly as tested on Dataset vX.Y with tool version/date. This prevents quiet history rewrites and helps readers understand drift.
Corrections Policy
We treat corrections like a scientific project: transparent, timestamped, and specific.
What we correct
- factual errors (pricing, features, limits)
- methodology mistakes (dataset labeling, preprocessing issues)
- reporting mistakes (wrong score, wrong screenshot, misinterpreted output)
What we do not do
- We do not remove negative findings because a vendor disagrees.
- We do not accept pay-to-improve-ranking arrangements.
- We do not publish guaranteed detection claims.
How corrections are handled
- Corrections are appended with a dated note (Updated on YYYY-MM-DD: …).
- Major corrections trigger a retest and a dataset/methodology version bump.
- Vendors can provide clarifications; we label them as vendor statements unless independently verified.
References (key sources that shape this methodology)
- OpenAI – New AI classifier for indicating AI-written text
- Stanford HAI – AI-Detectors Biased Against Non-Native English Writers
- Liang et al., Patterns (Cell) – GPT detectors are biased against non-native English writers
- Krishna et al., arXiv – Paraphrasing evades detectors of AI-generated text
- NIST AI 100-4 – Reducing Risks Posed by Synthetic Content
FAQ
How often is the methodology updated?
Quarterly, or sooner if major model/tool changes occur. We also publish trigger-based updates when detectors change engines, pricing, or core behavior.
Do you run real tests or rely on vendor claims?
Real tests first. Vendor claims are recorded for context, clearly labeled as vendor statements, and never substituted for reproducible results on our dataset.