Back to Research
Infrastructure Theory

The New Sorting Hat

AI Writing Detection as Classification Infrastructure

This essay argues that AI writing detectors are not accuracy problems waiting to be solved. They are classification infrastructure being installed across higher education, and infrastructure theory predicts that their deepest failures will compound, not self-correct, as the system becomes more embedded. It begins with a case. The case is not the argument. The case is what the argument explains.

In February, a judge found that Adelphi Universitys decision to punish a student for AI plagiarism was without valid basis and devoid of reason.3 Orion Newby, a freshman with documented learning differences who had worked with university-provided tutors throughout his writing process, submitted an essay on Christianity and Islam. His professor ran it through Turnitins AI detector. The tool flagged it as wholly AI-generated. Newby denied using AI.

The school gave him a zero and ordered him into an anti-plagiarism course, a remediation program for a category hed been wrongly sorted into, then warned that a second offense could mean suspension. When he appealed, the school declined to reexamine the allegations. It never interviewed his tutors. It never examined his drafts. It had a score.

Newbys family spent six figures on legal fees to prove what a conversation could have established for free.

His case is not unusual. Only the outcome is.

Yale

A French-born MBA student is suing after GPTZero flagged his exam answers. The university suspended him for a year. His lawsuit alleges discrimination against non-native English speakers.

University of Michigan

A student with OCD and generalized anxiety disorder is suing after professors interpreted her formal, structured writing style as evidence of AI generation.

Campuses nationwide

Students describe running their own work through multiple detectors before submitting, trying to pre-clear assignments they wrote themselves.

One recent graduate told NBC News that she and her classmates had a group chat called Academic Felons for Life.

The structural problem

The Thesis

The conversation about these tools has mostly been a technology debate: Do they work? Are they accurate enough? But these questions miss the structural problem entirely.

AI writing detectors are not technology products being evaluated by educators. They are classification infrastructure being installed across education in real time.56

IWhat Makes It Infrastructure

Infrastructure, in the precise sense that scholars of technology use the term, is not just the thing underneath the thing. It is a system that gets woven into institutional processes so thoroughly that it becomes invisible: taken for granted, difficult to question, and nearly impossible to remove.6 Infrastructure persists not because it is optimal but because the cost of replacing it exceeds the cost of living with its flaws.

Everything gets built on top of it, and eventually, removing the foundation means demolishing the building.

0+
institutions integrated with Turnitin

Detection scores feed into grading workflows, academic integrity databases, and student disciplinary records. Faculty build syllabi around the assumption that submitted work will be scanned. Policies reference detection thresholds.

When Purdue announced that Turnitins AI detection update would activate automatically, they disclosed that the university does not have the ability to turn it off.7

Purdue University, April 2023

The institution that licensed the tool could not control when the tool changed underneath them. That is not a product. That is infrastructure.

The cascade: what one integration connects to

Tap any node. Then tap Remove the tool.

Turnitin AI Detection
LMS
Grading
Integrity DB
Syllabi
Pedagogy

And infrastructure theory predicts that the tools three deepest problems will compound, not self-correct, as the system becomes more embedded.5

The three problems

IIThe Boundary Problem

AI detectors produce continuous probability scores: a percentage likelihood that text was generated by a machine.12 But institutions need categories: cheating or not cheating, flagged or clean. So they impose thresholds. Where you draw the line determines who gets accused.

The boundary problem: drag the threshold

Eight students. Two actually used AI. Move the threshold. Tap any student to see why they scored where they did.

0%Threshold: 20%100%
Student ANative English, creative style
3%
CLEAR
Student BADHD, highly structured writing
12%
CLEAR
Student CFrench-born, MBA program
18%
CLEAR
Student DESL, first-gen college student
24%
FALSE +
Student EUsed AI to brainstorm, wrote by hand
41%
FALSE +
Student FOCD, meticulous revision habits
67%
FALSE +
Student GPasted prompt, submitted output
89%
FLAGGED
Student HGenerated and submitted unchanged
95%
FLAGGED
5 flagged · 2 actual AI use · 3 false positives

Turnitin hides scores in this range. They show an asterisk instead.

Turnitin acknowledged the boundary problem seven weeks after launch. Their Chief Product Officer admitted they had discovered a higher incidence of false positives when less than 20% of AI writing is detected.2 She did not disclose the actual rate. Their fix was to display an asterisk instead of a number for scores between 1% and 19%.

They did not improve the tool in that range. They stopped showing you the score. The infrastructure kept running; they dimmed the dashboard light.

The number Turnitin advertises vs. the number you experience
<1%
False positive rate (document-level)

The chance of flagging a fully human-written paper as AI-generated. This is what Turnitin advertises.

The number the company advertises and the number the user experiences are measuring different things.

The loop

IIIThe Looping Effect

Students who know they will be scanned change their writing.8 Not to write better, but to write in ways that avoid triggering detection. They write less clearly, less concisely, less precisely, because clarity and predictability are exactly what the detectors flag.

Turnitins own documentation explains that it evaluates how statistically likely each word choice is given its context: language models choose optimal words, so human writers need to choose suboptimal ones to avoid suspicion.13 Better writing now looks more AI-like to the machine.

The looping effect
DetectionDeployedStudentsAdaptWritingDegradesHumanizersEmergeDetectorsRetrain

The system produces the adversarial behavior it was designed to prevent, then points to that behavior as justification for its own existence.

Faculty are adapting too. In-class handwritten exams replace take-home essays, not because they are pedagogically superior but because they are scan-proof. The classification system is restructuring pedagogy around itself.

150+
humanizer tools
charging up to $50/month
33.9M
monthly visits
to humanizer sites in a single month
$50/mo
price ceiling
students pay to rewrite their own work

The humanizer industry alone drew 33.9 million website visits in a single month.11 Students are paying for the privilege of making their own writing worse so that a machine will believe they wrote it.

The system is producing exactly the adversarial behavior it was designed to prevent, then pointing to that behavior as justification for its own existence.

Who it hits

IVDisparate Impact

Researchers at Stanford tested seven AI detectors and found that they classified over 61% of TOEFL essays written by non-native English speakers as AI-generated, while performing near-perfectly with essays by U.S.-born students.1

The reason is structural: non-native speakers tend toward simpler syntax, more predictable vocabulary, and more formulaic constructions. These are the exact features the detectors associate with machine output.

Stanford study: seven detectors tested
TOEFL essays flagged as AI0%
Non-native English speakers
U.S.-born student essays flagged0%
Native English speakers
0%
Black students falsely flagged
0%
White students falsely flagged

The detectors are not failing randomly. They are failing along the same lines that educational systems have always failed along.

Students with autism, ADHD, and other conditions that produce systematic or highly structured writing patterns face elevated false positive rates.4 The detectors are not failing randomly. They are failing along the same lines that educational systems have always failed along. And calling the result a technical measurement.

The deeper problem

VThe Category Is Incoherent

But beneath all three problems lies a deeper one: the category itself is incoherent. The detectors assume a binary: human-written or AI-generated. That binary is already obsolete.

Six students. One question the tool can answer. One it cannot.
STUDENT 1

Asked an LLM to brainstorm topic ideas, then researched and wrote the essay herself.

STUDENT 2

Drafted from scratch, asked an LLM to identify weak arguments, revised by hand.

STUDENT 3

Dictated rough ideas into a voice memo, fed the transcript to an LLM for an outline, then wrote from the outline.

STUDENT 4

Wrote a full draft, asked the LLM to suggest transitions, rejected most, adopted two.

STUDENT 5

Asked the LLM to generate a first draft, then substantially rewrote every paragraph in her own voice.

STUDENT 6

Pasted in a prompt and submitted the output unchanged.

The detector cannot distinguish between the student who used AI to think and the student who used AI to avoid thinking. The classification system was designed for a world where human and machine writing are discrete categories, and that world no longer exists.

What we are building is a system that punishes students for how their writing sounds rather than how they learned.

The psychiatrist Arthur Kleinman called this the category fallacy: importing a classification scheme into a context where it does not fit.9 The human/AI binary made a kind of sense before AI became a writing environment rather than a writing tool. It does not make sense now. But the infrastructure has already been built on top of it.

They know

The Company Knows

And the company that built the infrastructure knows it.10

This is the signature of mature infrastructure. The systems operators can describe its limitations in detail, publish documentation acknowledging those limitations, and build an entire support apparatus for managing the consequences of those limitations, all while continuing to sell and expand the system. The disclaimer does not constrain the infrastructure. The disclaimer is part of it.

What the infrastructure predicts

What Happens Next

Infrastructure theory predicts what happens next. The tools will become invisible in the technical sense: not hidden, but taken for granted. Unquestioned. That is what infrastructure scholars mean by transparent: not that you can see through it, but that you stop seeing it at all.

Infrastructure theory predicts

Switching costs compound

Every policy, every workflow, every database that references a detection score makes the system harder to remove.

Disparate impacts accumulate

False flags travel in student records, shaping institutional trajectories long after the detection score itself is forgotten.

The binary collapses

Writing is already becoming collaborative in ways no detector can parse. But the infrastructure persists because everything has been built on top of it.

We have seen this before. The diagnostic categories in the DSM have persisted for decades despite widespread acknowledgment of their scientific limitations, not because they are accurate but because insurance billing, treatment guidelines, and clinical training are all built on top of them. AI detection is on the same trajectory, compressed from decades into years because digital infrastructure installs faster than institutional infrastructure.

What would it take

Three Proposals

What would it mean to treat AI detection as an infrastructure design problem rather than a technology procurement decision? At minimum, three things.

Tap any proposal to see what happens when it is not adopted.

01Demographic transparency

If a detection tool cannot report its false positive rate by race, by language background, and by neurodiversity status, it should not be deployed in any context where those false positives carry consequences.

02Sunset clauses

Every AI detection contract should require re-evaluation on a fixed schedule (two years, not perpetuity) so institutions are forced to actively decide whether the infrastructure still serves them rather than allowing it to persist by default.

03A due process floor

An absolute prohibition on using detection scores as the sole basis for academic integrity proceedings. Turnitin’s own documentation already says this, in fine print, on every report, beneath the score that institutions are already treating as a verdict.

If you are a student

Document your writing process.

Drafts, outlines, revision histories, timestamped notes. If you are ever flagged, the process is the evidence. The score is not.

Know that a detection score is not a finding.

Turnitin{'’'}s own documentation says it should not be used as the sole basis for action against you. Quote it.

Do not pre-scan your own work.

Running your essay through multiple detectors before submitting teaches you to write for machines, not for readers. That is the looping effect, and you are inside it.

If you are a faculty member

Read the disclaimer at the bottom of the report you are about to act on.

If the tool{'’'}s own maker says the score should not be the sole basis for action, ask yourself what else you have.

Ask what you are assigning for.

If the answer is learning, and the tool cannot measure learning, the tool is measuring the wrong thing.

Track your own false positive patterns.

Which students get flagged? By what background? If you notice a pattern, the pattern is the data. Report it upward.

If you are an administrator

Ask your vendor for demographic false positive data.

If they cannot provide it, you are deploying a tool whose failure patterns you cannot see. That is a liability question, not a technology question.

Put a sunset clause in every detection contract.

Two years, not perpetuity. Force your institution to actively decide whether the infrastructure still serves you, rather than allowing it to persist because removing it is harder than keeping it.

Audit the secondary infrastructure.

How many hours of professional development, how many integrity hearings, how many student appeals are being generated by the tool? Count the cost of managing the tool{'’'}s failures against the cost of the tool itself.

Orion Newby got his grade back and his record expunged. It cost his family six figures and a federal court order.3

The infrastructure will sort the next student tomorrow morning, before anyone reads the disclaimer at the bottom of the screen.

Share this essay. Every section has a direct link.

Related research

The Sorting Machine

Special Education Classification as Infrastructure. The companion essay applying the same framework to IDEA disability categories.

References

13 sources · tap to expand
1.
Liang et al., 2023 (2023)
2.
Turnitin, 2023a (2023)
3.
Newby v. Adelphi, 2025 (2025)
4.
Weber-Wulff et al., 2023 (2023)
5.
Bowker & Star, 1999 (1999)
6.
Star, 1999 (1999)
7.
Purdue University, 2023 (2023)
8.
Hacking, 1995 (1995)
9.
Kleinman, 1988 (1988)
10.
Turnitin, 2023b (2023)
11.
SimilarWeb, 2024 (2024)
12.
Elkhatat, 2023 (2023)
13.
Sadasivan et al., 2023 (2023)
PDF

The New Sorting Hat

Full academic version with complete references and formatting

Download
jakelawrence.xyz
ResearchContact