RACE: Fine-Grained AI Text Detection That Distinguishes Human-Written, LLM-Polished, and Humanized AI Content
Available in: 中文
Existing AI text detectors only do binary (human vs AI) or ternary classification. This is insufficient because:
An ACL 2026 accepted paper introduces RACE (Rhetorical Analysis for Creator-Editor Modeling), a method that goes beyond binary "human vs AI" classification to detect fine-grained types of LLM involvement in text. This matters for policy enforcement, where LLM-polished human text and humanized LLM text carry different regulatory implications.
The Problem
Existing AI text detectors only do binary (human vs AI) or ternary classification. This is insufficient because:
- LLM-polished human text — Human writes, AI improves style. Common in professional writing.
- Humanized LLM text — AI generates, human edits. Common in content farms.
- Both trigger different policy consequences but existing detectors can't distinguish them.
The Four Classes
- Pure human text — Written entirely by humans
- LLM-generated text — Written entirely by LLMs
- LLM-polished human text — Human foundation, AI style editing
- Humanized LLM text — AI generation, human editing
How RACE Works
- Creator modeling — Uses Rhetorical Structure Theory to build a logic graph representing the original writer's foundation
- Editor modeling — Extracts Elementary Discourse Unit-level features to identify editing style
- Outperforms 12 baselines with low false alarm rates
Why It Matters
As AI writing tools become ubiquitous, regulation needs precision. A student using Grammarly-style polishing shouldn't be treated the same as someone submitting fully AI-generated content. RACE provides the granularity needed for fair enforcement.
← Previous: Predicting Student Video Behavior with Multimodal LLMs: 77 Million Events from 66 Online CoursesNext: AI Dark Patterns in Creative Writing: Study Reveals 91.7% Sycophancy Rate in LLM Writing Assistants →
0