Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

Contrary to widespread panic, research indicates that the percentage of students who self-report using AI to generate an entire assignment is only 10%. This figure has remained stable for cheating over the years, regardless of technology. Most students use AI to explain concepts or generate ideas, not to plagiarize wholesale.

Related Insights

The education system is fixated on preventing AI-assisted cheating, missing the larger point: AI is making the traditional "test" and its associated skills obsolete. The focus must shift from policing tools to a radical curriculum overhaul that prioritizes durable human skills like ethical judgment and creative problem-solving.

Data shows the vast majority (80%) of high school students use AI tools to explain concepts or brainstorm ideas. The rate of students admitting to cheating on entire assignments remains a consistent minority (~10%), suggesting AI is a new method for cheating, not a cause for a massive increase in it.

The recent surge in academic dishonesty is less about a moral decline and more a result of new AI tools making cheating easier to execute and significantly harder for educators to prove.

Students often use AI not out of laziness, but as a logical coping mechanism for an educational system prioritizing final grades over the learning process. Facing immense pressure from multiple courses and jobs, they see AI as a tool to produce a required "product" and survive, revealing a flaw in the system's incentives.

In response to AI making take-home assignments unreliable, universities are reverting to "old-school" assessment methods like in-class blue book exams, spontaneous writing sessions, and oral exams to ensure student work is authentic.

Relying on generative AI to produce assignments bypasses the effortful cognitive processes—like reflection and structuring arguments—that are essential for forming long-term memories. As a result, students who use tools like ChatGPT have very poor recall of the essays they submitted, defeating the purpose of the learning exercise.

Instead of just banning AI to prevent cheating, one school district experimented by increasing test frequency. This counterintuitively motivated students to use guided AI learning features to master the material, rather than just get homework answers, proving the need to rethink educational workflows.

While cheating is a concern, a more insidious danger is students using AI to bypass deep cognitive engagement. They can produce correct answers without retaining knowledge, creating a cumulative learning deficit that is difficult to detect and remedy.

Instead of policing AI use, a novel strategy is for teachers to show students what AI produces on an assignment and grade it as a 'B-'. This sets a clear baseline, reframing AI as a starting point and challenging students to use human creativity and critical thinking to achieve a higher grade.

Generative AI's appeal highlights a systemic issue in education. When grades—impacting financial aid and job prospects—are tied solely to finished products, students rationally use tools that shortcut the learning process to achieve the desired outcome under immense pressure from other life stressors.