After hundreds of conversations with teachers, administrators, and school districts, I keep seeing the same pattern: schools panicking about AI-generated essays and reaching for solutions that sacrifice the very skills writing assignments were designed to build.
The most common reaction? Retreating to timed, in-class essays. On the surface, it seems logical: if students can't access ChatGPT during a 45-minute class period, they can't cheat. Problem solved.
Except it's not solved. It's traded for a worse one.
The Point of the Assignment
Before asking how to prevent cheating, schools should be asking a more fundamental question: what is this assignment actually for?
Long-form, take-home essays exist because they develop cognitive skills that timed writing simply cannot. When a student sits with an idea over several days, drafts and restructures an argument, rearranges paragraphs, and revisits their thesis after sleeping on it, they are doing something fundamentally different from producing text under time pressure.
Research backs this up. A peer-reviewed study by Quitadamo and Kurtz (2007) compared students who completed writing-intensive lab treatments with those who took traditional quizzes in a general education biology course. The writing group significantly improved their critical thinking skills, specifically in analysis and inference, while the non-writing group did not (Quitadamo & Kurtz, 2007, CBE Life Sciences Education).
Writing is not just a way to demonstrate thinking. It is a mechanism for developing it. As Menary (2007) argues in Language Sciences, writing functions as a form of thinking itself, not merely a record of thought that happened elsewhere. The act of organizing, restructuring, and refining ideas on a page is where cognitive development happens.
A research summary from MSU Denver's Writing Center reinforces this, noting that writing facilitates deeper engagement with subject material because it "requires focusing of attention, planning and forethought, organization of one's thinking, and reflective thought" (MSU Denver, "Writing as a Thinking Tool").
Timed essays strip away most of this. As Olivia Henderson argued in a piece for UCSB's Interdisciplinary Humanities Center, timed essays "reward the ability to write and think quickly rather than write and think well." They don't allow students to conduct research while writing, and they don't evaluate the ability to gather information, digest it, and form independent thoughts about it. Henderson notes that "taking the time to think about the nuances of a text and craft a meaningful argument is a more difficult, but rewarding and critical, skill than writing quickly" (Henderson, 2022, UCSB IHC).
If in-class essays were a better pedagogical tool, we never would have moved to take-home writing in the first place.
The Productive Struggle That AI Erases
Cognitive psychologist Robert Bjork's research on "desirable difficulties" (1992, 2011) shows that learning conditions which create short-term challenges actually enhance long-term retention and transfer. Effortful learning outperforms easy learning because the struggle itself creates stronger neural connections. Information becomes deeply encoded through meaningful connections rather than surface-level repetition (Bjork & Bjork, 2011).
A 2025 report from Bellwether Education Partners, titled "Productive Struggle: How Artificial Intelligence Is Changing Learning, Effort, and Youth Development in Education," directly addresses this in the AI context. The report defines productive struggle as "the process of engaging with challenging tasks or problems that require effort, critical thinking, and persistence to solve," and notes that it typically includes running into obstacles, making mistakes, and experiencing discomfort while still working toward a solution (Bellwether, 2025).
"When does ease enable greater learning, and when is ease a shortcut with a hidden cost?"
This is exactly what many teachers I talk to are missing. The ones who have adopted the position that "AI is here to stay, just teach kids to use it" have a point, but they haven't followed the logic far enough. When an AI feedback tool explains what is wrong with a student's argument in real time, or generates alternative phrasings on demand, it removes the productive friction that makes writing valuable. The student gets a polished essay. They do not get the cognitive development that comes from wrestling with their own ideas.
Why Typed Essays Still Matter (More Than Handwritten Ones)
Some schools are going even further, switching to handwritten essays to eliminate the possibility of digital cheating entirely. This creates its own set of problems.
Yes, research from van der Meer and van der Weel (2024), covered in Scientific American, shows that handwriting is better for note-taking and memory retention because it forces selective processing. You can't write down everything a lecturer says, so you have to actively prioritize and consolidate information.
But note-taking and essay composition are fundamentally different cognitive tasks. For essays, typing has distinct advantages that matter for developing thinking skills:
Students who type produce significantly more text, which allows for more developed arguments (Thomas, Paine, & Price, 2003). More importantly, typed essays enable the iterative revision process—including creating drafts, rearranging paragraphs, restructuring arguments, and refining ideas across multiple sessions—that is central to developing higher-order thinking. This revision cycle (adding, rearranging, removing, and replacing) is where the deepest learning in writing actually occurs.
Digital composition also enables something handwriting cannot: a verifiable record of the writing process. When students type in a monitored environment, every keystroke, paste, deletion, and revision is captured. This creates a forensic trail that serves two purposes: it protects honest students by providing evidence of authentic work, and it gives teachers actionable data rather than gut feelings when something looks off.
This brings me to a problem I see constantly.
AI Detectors Don't Work. But Neither Does "Just Trust Your Gut."
Most teachers I talk to still believe they can tell when a student has used AI. We built a "Spot the AI Essay" game to test this, and the results consistently surprise educators. Students are getting increasingly sophisticated at humanizing AI-generated text, and the gap between what teachers think they can detect and what they actually catch is widening.
Meanwhile, AI detection tools have been shown to be unreliable. A landmark 2023 study by Weber-Wulff et al. found that even the best AI detection software is only correct about 80% of the time, and simple paraphrasing can drop detection rates dramatically. Stanford researchers found that these tools disproportionately flag non-native English speakers' writing. OpenAI shut down its own AI detection tool in 2023 due to its low accuracy rate (Weber-Wulff et al., 2023; Liang et al., 2023).
So if AI detection doesn't work, and teachers can't reliably spot AI writing by instinct, what's the alternative?
The answer is not detection. It's forensics.
Rather than asking "was this written by AI?" after the fact, the better approach is to monitor the writing process itself. When you can see that a student drafted an outline, wrote a rough first paragraph, deleted it, started over, gradually built their argument over three sessions, and revised their conclusion twice, you don't need an AI detector. The process IS the evidence.
And when you can see that a student opened a blank document, pasted 800 words at 2 AM, changed two commas, and submitted it, you don't need an AI detector for that either. You have concrete behavioral data to point to, which transforms an accusation into a conversation grounded in evidence.
This is fundamentally different from free playback tools like Draftback or Google Docs version history. Those tools require teachers to manually watch a replay of every student's writing session—an impossible task at scale when you have 150 students. They also require students to share their documents, creating opportunities for manipulation: a student could submit a plagiarized copy while sharing a clean document, forcing the teacher to manually compare versions. The approach doesn't just fail to scale; it adds work instead of removing it.
What schools need is a system that automatically captures and analyzes the writing process, categorizes different types of potential integrity issues (copy-paste, AI generation, collaboration where it wasn't permitted), and surfaces only the cases that warrant teacher attention—with the evidence already organized.
The Real Question Schools Should Be Asking
Everyone agrees that banning AI entirely is not the answer. But "embrace AI" has become a thought-terminating cliche that lets schools skip the harder conversation about what learning actually requires.
The administrators I talk to often discuss "embracing AI" without distinguishing between fundamentally different use cases. A teacher using AI to generate quiz questions is not the same as a student using AI to write their essay. The teacher is not the one who needs to learn the material. This distinction seems obvious, but it is routinely collapsed in school-level AI policy discussions. I've heard the argument "if teachers are allowed to use AI, why shouldn't students?" more times than I can count, and it misses the point entirely.
The real question is not "how do we catch cheaters?" or "how do we embrace AI?" It is: how do we preserve the cognitive processes that make students better thinkers?
That means protecting the conditions under which productive struggle happens. It means keeping long-form, typed, take-home essays as a core part of writing instruction rather than retreating to timed writes that test speed over depth. And it means monitoring the writing process itself rather than trying to reverse-engineer whether the final product was AI-generated.
The assignments that matter most are the ones where students sit with discomfort, wrestle with complexity, and come out the other side with ideas they built themselves. If we let the fear of AI cheating eliminate those assignments, we will have done more damage to student learning than AI ever could.
Written by The Checkmark Plagiarism Team.
Sources
- Bjork, R. A., & Bjork, E. L. (2011). Making things hard on yourself, but in a good way: Creating desirable difficulties to enhance learning. In M. A. Gernsbacher, R. W. Pew, L. M. Hough, & J. R. Pomerantz (Eds.), Psychology and the real world: Essays illustrating fundamental contributions to society. Worth Publishers.
- Bellwether Education Partners. (2025). Productive Struggle: How Artificial Intelligence Is Changing Learning, Effort, and Youth Development in Education. https://bellwether.org/publications/productive-struggle/
- Henderson, O. (2022). Time's Up!: Why We Should Stop Assigning Timed Essays. Interdisciplinary Humanities Center, University of California, Santa Barbara. https://ihc.ucsb.edu/times-up-why-we-should-stop-assigning-timed-essays/
- Liang, W., Yuksekgonul, M., Mao, Y., Wu, E., & Zou, J. (2023). GPT detectors are biased against non-native English writers. arXiv preprint arXiv:2304.02819.
- Menary, R. (2007). Writing as thinking. Language Sciences, 29, 621–632.
- MSU Denver Writing Center. Writing as a Thinking Tool. https://www.msudenver.edu/writing-center/faculty-resources/writing-as-a-thinking-tool/
- Quitadamo, I. J., & Kurtz, M. J. (2007). Learning to improve: Using writing to increase critical thinking performance in general education biology. CBE Life Sciences Education, 6(2), 140–154.
- Thomas, M. H., Paine, B. J., & Price, T. S. (2003). Referenced in Mogey, N., Paterson, J., Burk, J., & Purcell, M. (2010). Typing compared with handwriting for essay examinations at university. Alt-J, 18(1), 29–47.
- van der Meer, A., & van der Weel, R. (2024). Referenced in Scientific American: Why Writing by Hand Is Better for Memory and Learning.
- Weber-Wulff, D., et al. (2023). Testing of Detection Tools for AI-Generated Text. arXiv preprint arXiv:2306.15666.

