
Ailsa Ostovitz, left, and her mother, Stephanie Rizk, at their home in the Maryland suburbs of Washington, D.C. In mid-November, Rizk met with Ostovitz's teachers to discuss accusations that her daughter had used AI to do some of her schoolwork. Beck Harlan/NPR
High schoolers face plagiarism charges based on flawed detection tools showing probabilities under 50%, raising questions about academic integrity enforcement
Ailsa Ostovitz, a 17-year-old junior at Eleanor Roosevelt High School in Maryland, has been accused of using AI on three assignments across two classes this school year despite writing the work herself. "It's mentally exhausting because it's like I know this is my work," Ostovitz told NPR. "I know that this is my brain putting words and concepts onto paper for other people to comprehend."
The accusations stem from AI detection software that teachers increasingly use to identify AI-generated content. One screenshot shared with NPR showed a 30.76% probability that Ostovitz used AI on a writing assignment describing the music she listens to—a probability well below the 50% threshold that would indicate more likely than not AI involvement.
Detection Software Reliability Problems
AI detection tools claim to identify AI-generated content by analyzing writing patterns, word frequency, sentence structure, and other markers supposedly distinguishing human and machine text. However, these tools produce frequent false positives, flagging authentic student work as AI-generated based on statistical patterns that overlap between human and AI writing.
The 30.76% probability in Ostovitz's case illustrates the arbitrary nature of detection thresholds. Teachers and administrators often lack guidance about what probability level justifies academic integrity actions, leading to inconsistent enforcement where students face consequences for scores indicating their work is more likely human-written than AI-generated.
Studies have shown AI detection tools perform inconsistently, with accuracy rates varying by writing style, subject matter, and student demographics. Some research suggests the tools disproportionately flag non-native English speakers whose grammatical patterns differ from typical native speaker writing, creating potential discrimination issues.
Academic Consequences
Students accused of AI use face serious consequences including assignment failures, grade reductions, parent conferences, and academic integrity violations on permanent records that could affect college admissions. These punishments occur even when detection probabilities fall far below certainty thresholds used in other high-stakes decisions.
The mental and emotional toll on falsely accused students extends beyond grades. Ostovitz's description of the process as "mentally exhausting" reflects the stress of proving authorship of your own work, defending against algorithmic accusations, and facing institutional skepticism about your integrity.
Students often struggle to prove they wrote assignments themselves. Unlike plagiarism from human sources where original texts can be compared, disproving AI use requires demonstrating a negative—that AI wasn't involved. Draft histories, writing process documentation, and in-person writing samples provide some defense, but many students don't maintain such records for routine assignments.
Institutional Challenges
Schools face genuine challenges distinguishing authentic student work from AI-generated submissions as tools like ChatGPT produce increasingly sophisticated text. Teachers managing dozens or hundreds of students lack time to thoroughly evaluate every assignment's authenticity manually.
However, relying on flawed detection software as primary evidence for academic integrity violations creates injustice when false positives damage innocent students. The rush to implement technological solutions outpaced understanding of these tools' limitations and appropriate use cases.
Some educators argue that assignment design—emphasizing in-class writing, personal reflection, and original analysis rather than generic essays—provides better protection against AI misuse than detection software. This approach focuses on making AI less useful for assignments rather than trying to catch students after the fact.
Legal and Ethical Questions
The use of AI detection software raises questions about due process, evidence standards, and fairness in academic settings. When students face serious consequences based on probabilistic scores from proprietary algorithms they cannot examine or challenge, fundamental fairness concerns emerge.
Parents and students are beginning to push back against AI detection tools, demanding higher evidence standards before academic sanctions. Some have threatened or pursued legal action when false accusations damaged students' records or college prospects.
The Broader Context
The detection software controversy reflects broader tensions as AI writing tools become ubiquitous. Defining appropriate AI use in education, developing fair assessment methods, and distinguishing legitimate assistance from academic dishonesty remain unresolved challenges.
Rather than rushing to punish potential AI use, some educators advocate for teaching students how to use AI appropriately as a tool while maintaining original thinking and authentic learning. This approach acknowledges that students will encounter AI throughout their careers and education should prepare them for that reality.
For students like Ostovitz facing false accusations, the immediate concern is clearing their names and protecting their academic records from algorithmic errors with lasting consequences.



