Back in February, the plagiarism detection service Turnitin announced that it had developed an AI writing detector, and word on the street is that Turnitin will enable this new tool for campuses on April 4th. The purpose of this new tool is to analyze writing submitted by students and determine whether it's likely the writing was actually written by an AI text generation tool like ChatGPT. For instructors or institutions that have decided that student use of AI text generation tools counts as plagiarism or unauthorized aid or some other violation of academic integrity policies, Turnitin's new tool ostensibly provides a way to detect such violations.
Turnitin says that the new AI writing detector tool "focuses on accuracy--if we say there's AI writing, we're very sure there is." See this video with David Adamson, an AI scientist with Turnitin, for some details on their process. Turnitin reports a 1% false positive rate for the tool, which means that if you submit 100 samples of actual student writing, on average one of those samples will be (falsely) flagged as likely written by AI. That sounds like a pretty accurate tool, but please allow me to do a little math here in the newsletter and introduce you to the paradox of the false positive.
I'm not going to ask ChatGPT to explain the paradox of the false positive to you, although I bet it could do a pretty good job. Instead, let me quote from the young adult novel Little Brother by Cory Doctorow. This is the novel I ask my cryptography students to annotate collaboratively using Perusall, and I'm grateful to Doctorow for releasing the novel under a Creative Commons license so I can easily do so. Here's how he introduces the paradox of the false positive in Little Brother via his very opinionated narrator Marcus:
#####
#####
Doctorow goes on to apply this notion to the challenge of detecting terrorists. If the vast majority of people in a given population aren't terrorists, even if your hypothetical terrorist detector if very accurate, you'll end up with lots of false positives. And what might those false positives lead to? You can probably imagine, but I'll share the innocuous end of the spectrum: My friend Dave Stewart used to have a lot of problems getting on airplanes because some other guy named Dave Stewart was on a no-fly list somewhere. Of course, the other end of the spectrum involves the nastier parts of the United States legal system.
What does the paradox of the false positive imply for the use of AI writing detection tools like the new one from Turnitin? Let's suppose for a minute that you have three hundred student assignments you'll be grading this month, and let's further suppose that none of your students use ChatGPT or any other AI writing tool when writing these assignments. That 1% false positive rate from Turnitin's new tool means that about three of those three hundred assignments will be flagged as likely AI-written.
That's three assignments that you might feel compelled to investigate. How exactly will you go about confirming or denying Turnitin's assessment of the writing in question? Will you call those students into your office and share the Turnitin report with them? Will you ask them about their process for completing the assignment? Will you require them to somehow prove that they wrote the assignment without AI help? This gets messy pretty fast, and it's not likely to be easy for you or the innocent students.
To be fair, I am not someone who thinks Turnitin's tools should be rejected out of hand as unwarranted surveillance of students and their academic work. When teaching my writing seminar, I would regularly use Turnitin's "originality report" with students to help them understand better and worse ways of working with their sources. If the report flags a few sentences in a student's paper as matching some other source, that's a great opportunity to talk with that student about their decisions and methods for quoting or summarizing sources. (I adopted this approach from Vanderbilt chemistry instructor Michelle Sulikowski, whom I interviewed for a podcast episode way way back in 2008.) It's hard for me to imagine how an AI writing detection report from Turnitin could be used in similarly productive way.
The driver behind the paradox of the false positive is the ratio of students not using AI tools to students using AI tools. If that ratio is very high, that is, if most students aren't using these tools, then any false positive rate, even one as low as 1%, becomes problematic. The higher the ratio, the more problematic the false positive rate becomes. What percent of a given population of students are likely to use AI writing tools when they have been told not to do so? I don't know, but that's the probability that educators should be attending to and working to influence.
This is the part of the newsletter where I link to things that I find interesting in the hopes that you do, too.
If you found this newsletter useful, please forward it to a colleague who might like it! If you have thoughts on any of the topics above, please reach out via email, Twitter, Mastodon, or LinkedIn.
Consider supporting Intentional Teaching through Patreon. For just $3 US per month, you get access to bonus podcast content, Patreon-only teaching resources (like a guide to facilitating class sessions on Zoom), an archive of past newsletters, and a community of intentional educators. Patreon supporters also get 20% off my book, Intentional Tech: Principles to Guide the Use of Educational Technology in College Teaching, when they order from West Virginia University Press.
Welcome to the Intentional Teaching newsletter! I'm Derek Bruff, educator and author. The name of this newsletter is a reminder that we should be intentional in how we teach, but also in how we develop as teachers over time. I hope this newsletter will be a valuable part of your professional development as an educator.
Annotation and Learning with Remi Kalir It's one thing to pull a book off a shelf, highlight a passage, and make a note in the margin. That's annotation, and it can be a useful learning tool for an individual. It's another thing to share your annotations in a way that others can read and respond to. That's social annotation, and when I heard years ago about digital tools that would allow a class of students to collaboratively annotate a shared textbook, I thought, well, that's the killer app...
Structure Matters: Custom Chatbot Edition Many years ago when educators were seeing what they could do with Twitter in their teaching, I wrote a blog post noting that structured Twitter assignments for students seemed to work better than more open-ended invitations for students to use Twitter to post about course material. When we walked through my mom's house as it was being built, I couldn't help but take a photo of all those lines. Somewhat more recently, I started sharing the structured...
Students as Partners in Teaching about Generative AI Last year on the podcast, I talked with Pary Fassihi about the ways she was exploring and integrating the use of generative AI in the writing courses she teaches at Boston University. During that interview, Pary mentioned an AI affiliate program running out of the writing program at Boston University. This program involved matching undergraduate students—the AI Affiliates—with writing instructors, giving the AI Affiliate a role in...