We’re Training Students To Write Worse To Prove They’re Not Robots, And It’s Pushing Them To Use More AI
from the can-someone-ask-an-ai-about-incentives dept
About a year and a half ago, I wrote about my kid’s experience with an AI checker tool that was pre-installed on a school-issued Chromebook. The assignment had been to write an essay about Kurt Vonnegut’s Harrison Bergeron—a story about a dystopian society that enforces “equality” by handicapping anyone who excels—and the AI detection tool flagged the essay as “18% AI written.” The culprit? Using the word “devoid.” When the word was swapped out for “without,” the score magically dropped to 0%.
The irony of being forced to dumb down an essay about a story warning against the forced suppression of excellence was not lost on me. Or on my kid, who spent a frustrating afternoon removing words and testing sentences one at a time, trying to figure out what invisible tripwire the algorithm had set. The lesson the kid absorbed was clear: write less creatively, use simpler vocabulary, and don’t sound too good, because sounding good is now suspicious.
At the time, I worried this was going to become a much bigger problem. That the fear of AI “cheating” would create a culture that actively punished good writing and pushed students toward mediocrity. I was hoping I’d be wrong about that.
Turns out… I was not wrong.
Dadland Maye, a writing instructor who has taught at many universities, has published a piece in the Chronicle of Higher Education documenting exactly how this has played out across his classrooms—and it’s even worse than what I described. Because the AI detection regime hasn’t just pushed students to write worse. It has actively pushed students who never used AI to start using it.
This fall, a student told me she began using generative AI only after learning that stylistic features such as em dashes were rumored to trigger AI detectors. To protect herself from being flagged, she started running her writing through AI tools to see how it would register.
A student who was writing her own work, with her own words, started using AI tools defensively—not to cheat, but to make sure her own writing wouldn’t be accused of cheating. The tool designed to prevent AI use became the reason she started using AI.
This is the Cobra Effect in its purest form. The British colonial government in India offered a bounty for dead cobras to reduce the cobra population. People started breeding cobras to collect the bounty. When the government scrapped the program, the breeders released their now-worthless cobras, making the problem worse than before. AI detection tools are our cobra bounty. They were supposed to reduce AI use. Instead, they’re incentivizing it.
And this goes well beyond one student’s experience. Maye describes a pattern spreading across his classrooms:
One student, a native English speaker, had long been praised for writing above grade level. This semester, a transfer to a new college brought a new concern. Professors unfamiliar with her work would have no way of knowing that her confident voice had been earned. She turned to Google Gemini with a pointed inquiry about what raises red flags for college instructors. That inquiry opened a door. She learned how prompts shape outputs, when certain sentence patterns attract scrutiny, and ways in which stylistic confidence trigger doubt. The tool became a way to supplement coursework and clarify difficult material. Still, the practice felt wrong. “I feel like I’m cheating,” she told me, although the impulse that led her there had been defensive.
A student praised for years for being an exceptional writer now feels like a cheater because she had to learn how AI detection works in order to protect herself from being falsely accused. The surveillance apparatus has turned writing talent into a liability.
Then there’s this:
After being accused of using AI in a different course, another student came to me. The accusation was unfounded, yet the paper went ungraded. What followed unsettled me. “I feel like I have to stay abreast of the technology that placed me in that situation,” the student said, “so I can protect myself from it.” Protection took the form of immersion. Multiple AI subscriptions. Careful study of how detection works. A fluency in tools the student had never planned to use. The experience ended with a decision. Other professors would not be informed. “I don’t believe they will view me favorably.”
The false accusation resulted in the student subscribing to multiple AI services and studying how the detection systems work. Not because they wanted to cheat, but because they felt they had no other option for self-defense. And then they decided to keep quiet about it, because telling professors about their AI literacy would only invite more suspicion.
Look, I get it: some students are absolutely using AI to cheat, and that’s a real issue educators have to deal with. But the detection-first approach has created an incentive structure that’s almost perfectly backwards. Students who don’t use AI are punished for writing too well. Students who are falsely accused learn that the only defense is to become fluent in the very tools they’re accused of using. And the students savvy enough to actually cheat? They’re the ones best equipped to game the detectors. The tools aren’t catching the cheaters—they’re radicalizing the honest kids.
As Maye explains, this dynamic is especially brutal at open-access institutions like CUNY, where students already face enormous pressures:
At CUNY, many students work 20 to 40 hours a week. Many are multilingual. They encounter a different AI policy in nearly every course. When one professor bans AI entirely and another encourages its use, students learn to stay quiet rather than risk a misstep. The burden of inconsistency falls on them, and it takes a concrete form: time, revision, and self-surveillance. One student described spending hours rephrasing sentences that detectors flagged as AI-generated even though every word was original. “I revise and revise,” the student said. “It takes too much time.”
Just like my kid and the school-provided AI checker, Maye’s student spent a bunch of wasted time “revising” to avoid being flagged.
Students spending hours rewriting their own original work—work that they wrote—because an algorithm decided it sounded too much like a machine. That’s time taken away from studying, working, caring for family, or, you know, actually learning to write better.
Learning to revise is a key part of learning to write. But revisions should be done to serve the intent of the writing. Not to appease a sketchy bot checker.
What Maye articulates so well is that the damage here goes beyond false positives and wasted time. The deeper problem is what these tools teach students about writing:
Detection tools communicate, even when instructors do not, that writing is a performance to be managed rather than a practice to be developed. Students learn that style can count against them, and that fluency invites suspicion.
We are teaching an entire generation of students that the goal of writing is to sound sufficiently unremarkable! Not to express an original thought, develop an argument, find your voice, or communicate with clarity and power—but to produce text bland enough that a statistical model doesn’t flag it.
The word “devoid” is too risky. Em dashes are suspicious. Confident prose is a red flag.
My kid’s Harrison Bergeron experience was, in retrospect, a perfect preview of all of this. Vonnegut warned about a society that forces everyone down to the lowest common denominator by handicapping anyone who shows ability. And here we are, with AI detection tools functioning as the Handicapper General of student writing, punishing fluency, penalizing vocabulary, and training students to sound as mediocre as possible to avoid triggering an algorithm that can’t even tell the difference between a thoughtful essay and a ChatGPT output.
Maye eventually did the only sensible thing: he stopped playing the game.
Midway through the semester, I stopped requiring students to disclose their AI use. My syllabi had asked for transparency, yet the expectation had become incoherent. The boundary between using AI and navigating the internet had blurred beyond recognition. Asking students to document every encounter with the technology would have turned writing into an accounting exercise. I shifted my approach. I told students they could use AI for research and outlining, while drafting had to remain their own. I taught them how to prompt responsibly and how to recognize when a tool began replacing their thinking.
Rather than taking a “guilt-first” approach, he took one that dealt with reality and focused on what would actually be best for the learning environment: teach students to use the tools appropriately, not as a shortcut, and don’t start from a position of suspicion.
The atmosphere in my classroom changed. Students approached me after class to ask how to use these tools well. One wanted to know how to prompt for research without copying output. Another asked how to tell when a summary drifted too far from its source. These conversations were pedagogical in nature. They became possible only after AI use stopped functioning as a disclosure problem and began functioning as a subject of instruction.
Once the surveillance regime was lifted, students could actually learn. They asked genuine questions about how to use tools effectively and ethically. They engaged with the technology as a subject worth understanding rather than a minefield to navigate. The teacher-student relationship shifted from adversarial to educational, which is, you know, kind of the whole point of school.
That line Maye uses: “these conversations were pedagogical in nature” keeps sticking in my brain. The fear of AI undermining teaching made it impossible to teach. Getting past that fear brought back the pedagogy. Incredible.
This piece should be required reading for every educator thinking that “catching” students using AI is the most important thing.
As Maye discovered through painful experience, the answer is to stop treating AI as a policing problem and start treating it as an educational one. Teach students how to write. Teach them how to think critically about AI tools. Teach them when those tools are helpful, when they’re harmful, and when they’re a crutch. And for the love of all that is good, stop deploying detection tools that punish good writers and push everyone toward a bland, algorithmic mean.
We are, quite literally, limiting our students’ writing to satisfy a machine that can’t tell the difference. Vonnegut would have had a field day.
Filed Under: ai, ai detection, cheating, dadland maye, students


Comments on “We’re Training Students To Write Worse To Prove They’re Not Robots, And It’s Pushing Them To Use More AI”
anyone have a link to the full text of the source?
Re:
You can find it here:
https://archive.is/j2jbW
Re: Re:
Ironically, only if the person you’re replying to can prove they’re not a robot. I’m not gonna bother, so I can’t; the same goes for the Cobra Effect link.
An experiment somebody should try: run every piece of official communication from your school through those alleged “AI” detectors.
Re:
The college administration has been sending out anti-union propaganda emails from the college president right before the employees strike for a new contract and the emails are full of LLM “assisted” text but also notably wildly incorrect math that purports the massive increases the employees are being offered. They moved a decimal on accident and the increase claimed is actually the amount the employees want, but aren’t being offered.
But hey, making more than $300k a year is hard to do without the help of an LLM to talk to the peasants for you!
Re:
This is a good idea where students can run the AI detector on arbitrary writing. But in a school that uses AI detectors the way the schools I attended used plagiarism detectors, the detector runs only whenever the student submits something for an assignment. Options for the student to follow your idea include:
Problem is: if ai wasn’t being abused by bad actors and/or companies, they’re wouldn’t be the need for so detection.
The word “devoid” is too risky. Em dashes are suspicious. Confident prose is a red flag.
Just my opinion:
I’m more concerned that even bland words will be suspicious and eventually students will start writing like robots to a point that it will hurt them later in life.
Re:
Sure it is abused by bad actors but how impactful is that? Why punish people for the bad actions of others?
This is also a lesson on why you can’t leave critical thinking to a computer or algorithm.
Re:
If people write like robots long enough, the robots will start writing like people writing like robots.
I have definitely not seen any signs of it harming students with strong vocabularies, who have distinct writing styles. Where I HAVE seen it appearing to become a problem is precisely where you would expect. Systems designed to trend towards averages of their training material, produce distinctly average output, and so the students I see hitting issues from it are the most normal, average ones.
The poor ones, it’s obvious when they used it, because they are the ones who don’t know when questioned what words in their essays mean, and cannot summarise what they were trying to communicate with it in their own words.
The obviously good ones, their writing clearly stands out as not being written with assistance because they tend to be the ones with solidly referenced work who can, when asked, describe the content of the sources they used.
But for anyone who turns in work that definitely qualifies for a passing grade, but perhaps is written in a less confident manner, so hews heavily towards adherence to broader style guides, has a less showy lexicon, etc. That is the kind of writing which generated stuff always reminds me of when I see it.
But eventually we’ll get to the common good which AI will provide. Right?
So close, almost there, man.
Re:
I’m sure it’ll be here any day now!
Any day now!
… Any day now…
…
“The lesson the kid absorbed was clear: write less creatively, use simpler vocabulary, and don’t sound too good, because sounding good is now suspicious.”
Sounding good has always been suspicious.
“Cheating” shouldn’t even be penalized. It’s a non-issue. If the point of education is to teach students, they’re only cheating themselves.
Re:
I’m not sure that approach makes much sense for fields whose professionals can hurt people if they aren’t truly competent (medicine and arrchitecture come to mind).
Re: Re:
The licensing of architects is a relatively recent thing, and while it came about after people blamed architects for deaths, it’s not obvious whether that was really a sensible reaction. Stuff like locked fire exits in theaters could just as well be handled with inspections, and perhaps engineer approvals for certain types of occupancy.
On the other hand, there are forms of quack medicine such as chiropractic for which states explicitly allow people to call themselves doctors, sometimes with certification requirements and everything.
As for actual medicine, I consider experience a lot more important than schooling. Is it really so important to have a doctor who can write a good essay? I hope computer assistance won’t let them fake their way through a residency.
“We’re Training Students To Write Worse ”
… who is the ‘We’ in that indictment ?
and how widespread is the alleged problem
You see these patterns in the cycles of technology development where you go from the wild west, free range, full of possibilities stage to consolidation and legislation and eventually full enshittification and wealth-funded corporate dominance. And there are victims along the way who pay the price for being in front of the train. This happened with computers and networking and the hacking freakouts of the FBI (i.e. Bruce Sterling’s Hacker Crackdown), with copyright enforcement and impoverished teens being prosecuted for piracy (Napster, Limewire, infinite lawsuits), etc. College instructors have used demanded proctored exams that require webcams that record eye movement to detect cheating and invade privacy. Meanwhile, douche DOGE dropouts use ChatGPT for bigotry prompts to cut funding to breast cancer research.
Again again, tech is a tool and tools will be used by the poor for opportunities and often as the only, insufficient, inhuman resource for a modicum of care and support while the one percent uses it to dehumanize, commodify, and dominate for profit and power.
It is also an example of Goodhart’s Law: when a measure (an AI checker) becomes a target (output of the AI checker determines your grade), it ceases to be a good measure (students will focus on the output of the AI checker).
This is a risk that is inherent to using any form of grading. Getting a passing grade becomes the student’s primary goal, overriding the learning experience.
—
Re: ??
Good comment
Re:
The bots can take the em-dash from my cold, dead hands.
People learn to write, by reading, and learning to emulate what they read.
AIs learn to write, by reading, and learning to emulate what they read.
If your detection tool flags something, that doesn’t mean it was written by AI, it means that whoever wrote it learned from some of the same sources the AI did.
Ironically, it sounds like in trying to flush out artificial intelligence, these tools are only succeeding in encouraging the artificial and punishing intelligence.
Vonnegut wouldn’t have a field day, he’d go have a breakfast of champions.
Coding and other uses
As a software developer I’ve let LLM’s write some code and in many cases it produces good code in far less time than it would take me to type it in. You have to be vigilant though, review the code and regularly guide the AI to write the code the way you like it. Sometimes for performance, but most of the time to get maintainable code. I do see LLM’s used more in the future as “coding aids”.
I also see no reason why one would not use AI for other routine tasks. Why should one draft a formal “I’ll end my cable TV service at the end of next month” notice all by hand if one can have an LLM write it so that you only need to print and sign it after a quick review?
LLM’s can also create summaries of documents. You can ask them to review dozens of documents and summarize what they say about a specific topic before deciding which three of the documents to study in detail. Possibilities are there and it would be wise to teach students on how to make proper use.
I heard a manager at my company say: “You can give me an answer that an AI gave you, but you still take responsibility for its correctness!”
Re: Tool use
Yes, this is how we should be using AI… as a tool to achieve more. It doesn’t replace us, it assists us.
The policy at my place of work is that AI tool usage is acceptable (though not mandatory), but the employee still takes 100% responsibility for the output in their name. So if you let AI fck up… you fcked up, and YOU will find out.
Like you, MathFox, I’ve used AI to save typing time, and make suggestions on how to solve problems, including diagnosing some of the more esoteric (to put it politely) error messages from XML/XSLT processors (for example). But all the code from my computer is still committed in my name, and I’m the one that is in my periodic reviews, not AI.
Re:
But was typing ever the bottleneck? For me, it’s always the thinking. When I end up with problems to debug, it’s almost always because I didn’t think enough in the first place. I started coding before I figured out and wrote down the exact behavior I wanted. That’s the same reason I find it hard to review other people’s code: if I can hardly understand what I’m thinking, how will I understand what they were thinking?
So, I have serious reservations about a system where all code will be “someone else’s” code—and, worse, someone whose behavior I’ll never be able to predict, whereas I do eventually get some sense of the type of faults I can expect from each human co-worker. I’m also reminded of an ex-Soviet co-worker once talking about how “coder” was a distinct and lesser job than “programmer” over there; the LLM might be able to code, but can it program?
I’m more optimistic about using such systems to help with code review. Then again, we’ve all had to silence bogus computer-generated warnings and sometimes suggestions. From compilers, maybe from linters and more complex analysis tools, and so it’s an open question whether these things will be like Clippit or like a thoughtful and meticulous co-worker.
Mostly agree
Not all of us who teach first-year college writing are using AI detectors or policing cheating. In fact, AI can be a matter of equity. Most students do not have English professors for parents like my kids had. Or have the money to pay someone to write or proofread papers for them.
The only answer I have is a kind of “ungrading.” If students are worried about maintaining high grades, then I take that worry away as best I can. I’ve stopped using rubrics. I’ve mostly stopped requiring assignments to be a certain number of pages. I make intensive comments on what they write so they see I am making an effort to teach them. I tell them when I like a sentence they wrote. It’s not easy to do when class sizes are increasing, but I do my best and I ask them to give me a chance to teach them something.
Shockingly, I am in full agreement with Masnick on one of his AI posts. This is a different context, but there’s also a problem with self-proclaimed experts who swear that they can teach people how to spot synthetic content, and they’ve turned a moral panic into an inquisition. I’m a feature writer and also operate a YouTube channel, and for both I now have to carefully structure what I do so that the witch hunters don’t harass me and the gatekeepers don’t throw me out.
Will this make you reconsider publishing pro-regurgitation-engine push pieces by grifters?
Probably not, I’m betting, especially if they can camouflage it with a smidgen of ‘copyright delenda est’.
Meanwhile, we have a study outta OpenAI saying a. hallucinations are inextricable from how LLMs work and cannot be solved by additional training or better models and b. that the rate of hallucinations increases as the models have gotten more sophisticated. It turns out that making the lying machine better doesn’t make it more trurhful; it makes it better at lying.
Presumably that’s why people whose livelihoods are predicated on lying, such as tech CEOs and media folk, think it’s so impressive.
AI detectors
English professor here.
I am not allowed to use an AI detector to accuse students of using AI and neither are most faculty in most places in academia; AI detectors are pretty useless and most people who teach know this (or should). Even if I were using an AI detector, an 18% detection rate would not even merit a second look. This post would probably get about that same percentage.
I have students who have AI write their papers on an eighth grade level; it still reads like AI. But unless I can prove they used AI, there is nothing I can do about it but grade what they submit. Most students are not good at writing AI prompts and so their submissions often do not follow instructions or there is actual academic misconduct, all of which I can grade the students for without even mentioning AI.
The best way to catch students using AI is to ask them to talk about their work. Most of them don’t even read the papers AI writes for them and they have no idea what “their” papers say.
AI Pain
As a regular commenter on Bezos’ Washington Post trying to work out which words or phrases the recently introduced AI moderator objects to, I feel your kid’s frustration. Misspelling and splitting words where they shouldn’t be split to get round it is common there now. Luckily unlike school, subscribing isn’t compulsory and the scores won’t dog my life.
More training data?
So do all those essays being checked by the “AI detectors” become part of the training corpus for the AIs?
Harrison Bergeron
This is an aside, but if you haven’t seen the movie adaptation of Harrison Bergeron you should. Pay particular attention to the US president in the movie, he’ll remind you of someone.