intermediate
5 min read
Tuesday, April 7, 2026

Beyond Binary: Unmasking the Creator and Editor in LLM-Generated Text

The lines between human and AI text are blurring, creating a headache for developers building responsible AI. This paper introduces a groundbreaking four-class detection method that distinguishes not just *if* text is AI-generated, but *how* – by identifying the distinct signatures of the creator and editor, whether human or LLM. Get ready to build more sophisticated content moderation, compliance, and review systems.

Original paper: 2604.04932v1
Authors:Yang LiQiang ShengZhengjia WangYehan YangDanding Wang+1 more

Key Takeaways

  • 1. Traditional LLM detection methods (binary/ternary) are insufficient for the nuanced policy and ethical challenges of modern AI text.
  • 2. RACE introduces a rigorous four-class detection system: Pure Human (HH), LLM-Polished Human (HL), Humanized LLM (LH), and Pure LLM (LL).
  • 3. It uniquely identifies the 'creator's signature' using Rhetorical Structure Theory (RST) for logical structure and the 'editor's signature' using Elementary Discourse Unit (EDU)-level features for stylistic choices.
  • 4. This dual-role modeling allows for precise identification of how human and LLM contributions combine, even in complex collaborative scenarios.
  • 5. Developers can leverage this research to build more sophisticated content moderation, academic integrity, compliance, and developer productivity tools, fostering responsible AI use.

The Paper in 60 Seconds

The Problem: Current LLM detectors are largely binary (pure human vs. pure AI) or, at best, ternary (adding a 'collaborative' category). This isn't enough for the real world. Policy decisions often hinge on *how* AI was used. Is it a human's idea polished by an LLM, or an LLM's idea humanized for deception?

The Solution: The paper introduces RACE (Rhetorical Analysis for Creator-Editor Modeling), a novel method for fine-grained LLM detection across a rigorous four-class setting:

1.Pure Human (HH): Human creator, human editor.
2.LLM-Polished Human (HL): Human creator, LLM editor.
3.Humanized LLM (LH): LLM creator, human editor.
4.Pure LLM (LL): LLM creator, LLM editor.

How it Works: RACE is groundbreaking because it separates the *creator's* signature from the *editor's* signature. It uses Rhetorical Structure Theory (RST) to build a logic graph that reveals the creator's foundational structure, and extracts Elementary Discourse Unit (EDU)-level features to identify the editor's stylistic choices.

The Impact: RACE outperforms existing baselines, offering a policy-aligned solution for LLM regulation with significantly lower false alarms. This means developers can build tools that understand the true provenance of text.

Why This Matters for Developers and AI Builders

The era of simple 'AI-or-not' detection is rapidly coming to an end. As Large Language Models (LLMs) become seamlessly integrated into every facet of content creation – from drafting emails and generating code to summarizing reports and crafting marketing copy – the line between human and machine output is not just blurring; it's becoming a complex, multi-layered tapestry. For developers and AI builders, this complexity isn't just an academic curiosity; it's a critical challenge demanding sophisticated solutions.

Consider the implications:

A human writes a compelling blog post, then uses an LLM to polish it for clarity and conciseness (HL). This is generally acceptable, perhaps even encouraged.
An LLM generates an entire news article, which a human then subtly edits to remove AI-isms and inject a personal touch, making it appear human-originated (LH). This could be highly deceptive and problematic.

Both scenarios involve human and LLM collaboration, but their provenance and intent are vastly different, triggering distinct policy consequences in areas like academic integrity, journalistic ethics, legal compliance, and social media moderation. Building responsible AI means understanding not just *if* text is AI-generated, but *how* it was generated and by whom. This research provides the framework for building the next generation of tools that can make these crucial distinctions, fostering trust, preventing misuse, and enabling targeted interventions.

Beyond the Binary: The Four-Class Revolution

Traditional LLM detection methods often fall short because they operate on an oversimplified understanding of human-AI collaboration. A binary classifier can only tell you 'human' or 'AI.' A ternary system might add a 'mixed' or 'collaborative' category, but even this lumps together fundamentally different scenarios. The key insight of this paper is that LLM-polished human text (HL) and humanized LLM text (LH) are not the same, and treating them as such leads to flawed policies and ineffective moderation.

RACE's four-class setting provides the necessary granularity:

Pure Human (HH): The gold standard of human-only creation and editing.
LLM-Polished Human (HL): A human's core idea, enhanced and refined by an LLM. Think of it as an advanced spell-checker and grammar assistant on steroids.
Humanized LLM (LH): An LLM's initial output, carefully tweaked by a human to mask its AI origins. This category is often the focus of concern regarding deception.
Pure LLM (LL): Content entirely generated and (potentially) edited by an LLM, with no human intervention.

This nuanced classification empowers developers to build systems that reflect the true complexity of modern content creation and respond appropriately to each scenario.

RACE: Unpacking Creator and Editor Signatures

The brilliance of RACE lies in its ability to disentangle the distinct 'signatures' of the creator and the editor, even when both are involved and one of them is an LLM. The core hypothesis is that LLMs and humans leave different traces when they *originate* ideas (the creator role) versus when they *refine* existing text (the editor role).

Creator's Signature via Rhetorical Structure Theory (RST)

To identify the creator's foundation, RACE leverages Rhetorical Structure Theory (RST). RST is a linguistic framework that analyzes text organization by identifying rhetorical relations (e.g., cause-effect, elaboration, contrast, background) between segments of text. Imagine a text as a tree structure, where leaves are elementary units and branches represent how these units logically connect to form larger ideas.

How RACE uses RST: It constructs a logic graph based on the rhetorical structure of the text. Humans, when creating, tend to exhibit more complex, varied, and perhaps less predictable rhetorical patterns. LLMs, especially earlier generations, might follow more standardized or repetitive logical flows, particularly when generating from scratch. By analyzing this structural blueprint, RACE can infer the likely creator.

Editor's Signature via Elementary Discourse Units (EDUs)

To identify the editor's style, RACE focuses on Elementary Discourse Units (EDUs). EDUs are the smallest meaningful units of discourse, roughly corresponding to clauses. The editor's role is often about refining language, improving flow, and adjusting tone – changes that manifest at a finer textual granularity.

How RACE uses EDUs: It extracts a rich set of EDU-level features that characterize stylistic choices. These can include lexical diversity (vocabulary richness), syntactic complexity (sentence structure variations), specific word patterns, punctuation habits, coherence markers, and even common 'AI-isms' or 'human-isms.' An LLM polishing human text might standardize language, improve fluency, or make it more concise. A human editing LLM text might inject more personal flair, correct awkward phrasing, or break up monotonous sentence structures.

By combining these two distinct analytical layers, RACE can accurately untangle who originated the core message and who shaped its final presentation, providing unparalleled insights into text provenance.

What Can You BUILD with RACE? Practical Applications for Developers

The power of RACE extends far beyond academic research. For developers, this fine-grained detection capability opens up a new frontier for building more intelligent, ethical, and robust AI-powered applications.

Advanced Content Moderation Platforms: Imagine a social media platform that doesn't just block 'AI spam,' but can differentiate. Content identified as HL (human-created, LLM-polished) might be allowed to pass, perhaps with a voluntary disclosure. Content identified as LH (LLM-created, humanized to evade detection) could be automatically flagged for human review or even removed. This prevents the spread of sophisticated misinformation while allowing users to leverage AI for legitimate purposes.
Enhanced Academic Integrity Tools: Move beyond simple plagiarism checks. Educational platforms can integrate RACE to provide more nuanced feedback. A student using an LLM for proofreading and grammar correction (HL) could be guided on ethical AI use, while a student submitting an essay primarily generated by an LLM and then minimally tweaked (LH) would face appropriate academic consequences. This fosters responsible AI literacy.
Legal & Compliance Tech: In industries where document authenticity and human oversight are paramount (e.g., legal, finance, healthcare), RACE can be integrated into document management systems. It could verify that critical legal briefs, financial reports, or regulatory filings are human-originated (HH or HL), even if LLMs were used for drafting support. This mitigates risks associated with AI-generated errors, misrepresentations, or even deepfake text that could have serious legal ramifications.
Developer Productivity & DevTools: Consider AI-assisted code generation or documentation tools. Integrating RACE could allow development teams to audit the provenance of generated code snippets or technical documentation. Knowing if a complex function was primarily AI-created and then human-tweaked (LH) versus human-created and then LLM-optimized (HL) impacts trust, debugging strategies, and intellectual property attribution. This could be integrated into CI/CD pipelines to flag potentially risky AI-generated content.
Customer Service & Chatbot Quality Assurance: For companies using LLMs in customer service, RACE could be used for quality assurance. It could evaluate if 'human agent responses' were genuinely human (HH), LLM-assisted for efficiency (HL), or if an LLM was primarily generating responses that a human then lightly edited (LH) to mimic empathy. This helps maintain service quality and ensures transparency with customers.

The Future is Fine-Grained

As LLMs continue to evolve and become indistinguishable from human writers, the need for sophisticated, fine-grained detection will only grow. RACE offers a powerful, policy-aligned solution that moves us beyond simplistic binary judgments. By understanding the distinct roles of the creator and editor, whether human or AI, developers are now equipped to build the next generation of transparent, trustworthy, and ethically sound AI systems. The future of AI content is not just about generation; it's about intelligent, nuanced understanding of its origins.

Cross-Industry Applications

DE

DevTools/SaaS

Integrate RACE into AI-assisted code generation, documentation, and content creation platforms to audit the provenance of generated assets.

Ensures human oversight on critical intellectual property, improves trust in AI-generated code, and helps enforce responsible AI development practices.

CO

Content Platforms (Social Media, News, Publishing)

Implement fine-grained content moderation systems that distinguish between LLM-polished human content (potentially acceptable) and humanized LLM content (potentially deceptive or spam).

Significantly reduces the spread of sophisticated misinformation and deepfake text, while allowing for beneficial AI assistance, fostering a healthier and more transparent online environment.

FI

Finance/Legal

Utilize RACE in document management and auditing systems to verify the human origin of sensitive financial reports, legal briefs, or regulatory compliance documents, even if LLMs were used for drafting or editing.

Mitigates risks associated with AI-generated errors or misrepresentations in high-stakes environments, ensuring accountability and adherence to critical regulations.

ED

Education Technology

Develop next-generation academic integrity tools that can differentiate between a student using an LLM for proofreading (HL) versus having an LLM write the core content (LH).

Provides educators with a more nuanced understanding of student work, promotes ethical AI use in learning, and helps maintain academic standards in an AI-assisted world.