Towards Transparency: How Can We Distinguish AI from Human Text Going Forward?
The Future of Writing: A Symposium for Teachers
University of Southern California
Anna Mills, May 1, 2023
Licensed CC BY NC 4.0
Welcome!
I like to approach this in a spirit of inquiry. It’s a complex topic, and I don’t claim to have the answers.�
Feel free to ask questions in the Q&A as we go, and I’ll try to pause periodically to respond.
Slides (open for commenting): https://bit.ly/TowardsTransparencyGPT�
What to expect
Presentation by Anna Mills, licensed CC BY NC 4.0.
Let’s take as a starting point the idea that language models like ChatGPT can now produce�passable prose of many kinds
Not copied verbatim (at least not usually)
Different results are possible each time from the same prompt
Flexible on genre and style
Often sounds plausible
Presentation by Anna Mills, licensed CC BY NC 4.0.
Are there settings where we want to know if text is AI-generated? (For me there are)
Presentation by Anna Mills, licensed CC BY NC 4.0.
Let’s think not just about student writing but about writing more broadly
Presentation by Anna Mills, licensed CC BY NC 4.0.
We might want to determine which of the following categories each kind of text falls into, depending on the situation
Presentation by Anna Mills, licensed CC BY NC 4.0.
Maybe knowing if text is AI generated is even a right? From the Blueprint for an AI Bill of Rights:
How important is it to know if the following kinds of text are AI generated?
Rate the different kinds on Mentimeter!
Possible reasons to want to know
Presentation by Anna Mills, licensed CC BY NC 4.0.
Textpocalypse, anyone?
“What if, in the end, we are done in not by intercontinental ballistic missiles or climate change, not by microscopic pathogens or a mountain-size meteor, but by … text? Simple, plain, unadorned text, but in quantities so immense as to be all but unimaginable—a tsunami of text swept into a self-perpetuating cataract of content that makes it functionally impossible to reliably communicate in any digital setting?”
– Matthew Kirschenbaum, “Prepare for the Textpocalypse,” The Atlantic, March 8, 2023
Presentation by Anna Mills, licensed CC BY NC 4.0.
Kirschenbaum forsees “a debilitating amalgamation of human and machine authorship. ”
“Think of it as an ongoing planetary spam event, but unlike spam—for which we have more or less effective safeguards—there may prove to be no reliable way of flagging and filtering the next generation of machine-made text.”
– Matthew Kirschenbaum, “Prepare for the Textpocalypse,” The Atlantic, March 8, 2023
Presentation by Anna Mills, licensed CC BY NC 4.0.
Just today, AI scientist Geoffrey Hinton left Google warning of something like a “textpocalypse”
“"His immediate concern is that the internet will be flooded with false photos, videos and text, and the average person will “not be able to know what is true anymore.”"
“‘The Godfather of A.I.’ Leaves Google and Warns of Danger Ahead” The New York Times, May 1, 2023
Presentation by Anna Mills, licensed CC BY NC 4.0.
AI text is based on human sources that should be credited: “Data Dignity”
“At some point in the past, a real person created an illustration that was input as data into the model, and, in combination with contributions from other people, this was transformed into a fresh image. Big-model A.I. is made of people—and the way to open the black box is to reveal them.
This concept, which I’ve contributed to developing, is usually called “data dignity.” ”
Jaron Lanier, “There Is No AI,” The New Yorker, April 20, 2023
Presentation by Anna Mills, licensed CC BY NC 4.0.
Text helps us build a human relationship if we know it comes from a human
Presentation by Anna Mills, licensed CC BY NC 4.0.
For example, we may want to get to know someone via text
Presentation by Anna Mills, licensed CC BY NC 4.0.
If we value writing as a way to assess student learning, we will want to distinguish AI text from student writing
Even with ungrading approaches, we are assessing some baseline competence necessary for passing
Presentation by Anna Mills, licensed CC BY NC 4.0.
Doesn’t “embracing” student use of AI text generation also require some non-AI student work we can assess?
Presentation by Anna Mills, licensed CC BY NC 4.0.
I designed a module on Canvas Commons that asks students to reflect on the differences between ChatGPT’s critical assessment and a human-written assessment
Presentation by Anna Mills, licensed CC BY NC 4.0.
What did ChatGPT miss? What did its output get right?
How do those observations match what we learned about how language models work?
How might the sample essay have turned out if the student had started with the ChatGPT output and revised from there?
What lessons do you draw from this comparison?
A Canvas Commons module
But anyone could use ChatGPT to comment on ChatGPT output.
If you (always or sometimes) want to know if a text was AI-generated, why? What reason or reasons do you find most compelling?
Presentation by Anna Mills, licensed CC BY NC 4.0.
If we do want to distinguish AI from human text, at least in some circumstances, how might we do it?
Some consider this futile, but let’s consider
Dr. Sarah Elaine Eaton:
“Hybrid writing…will be the norm. Trying to determine where the human ends and where the artificial intelligence begins is pointless and futile.”
“6 Tenets of Postplagiarism: Writing in the Age of Artificial Intelligence”
Presentation by Anna Mills, licensed CC BY NC 4.0.
Maybe we can tell if the text is AI? (Use our noses?)
Presentation by Anna Mills, licensed CC BY NC 4.0.
What would be the giveaways?
I’m skeptical. How many of us can tell, how much of the time, and for how long?
I’m not as confident as Maha Bali and Lauren Goodlad that I can tell. Sometimes student writing has formulaic features. And ChatGPT can be “voicier.”
Can we be sure language models will stay mostly generic, bland, superficial in their responses?
Can we be sure they won’t be able to imitate individual styles and levels of proficiency with Standard English convincingly?
Presentation by Anna Mills, licensed CC BY NC 4.0.
Trust- and motivation-
based approaches
Let’s go back to why we assign writing. Not for the product. For the thinking process.
“A fundamental tenet of Writing Across the Curriculum is that writing is a mode of learning. Students develop understanding and insights through the act of writing. Rather than writing simply being a matter of presenting existing information or furnishing products for the purpose of testing or grading, writing is a fundamental means to create deep learning and foster cognitive development.”
Presentation by Anna Mills, licensed CC BY NC 4.0.
First and foremost, emphasize purpose and engagement
Presentation by Anna Mills, licensed CC BY NC 4.0.
Discuss the ethics of transparency with students
Presentation by Anna Mills, licensed CC BY NC 4.0.
“I also feel like our relationship with AI will become entangled and complex, where it will be difficult to determine where our agency begins and where the machine’s influence on us is stronger – but I sense that it is important for students’ metacognition that they remain aware of how they’ve integrated AI into their thinking��… this is why I suggest an approach of “transparency“, not just of students disclosing where they got some ideas/text from in the sense of attribution, but also of reflecting on where they used AI in their process and why that was helpful or how it needed tweaking, etc..”
Presentation by Anna Mills, licensed CC BY NC 4.0.
Trust may be one good path and may suffice in many situations
But will it in others? In some settings, incentives and competitive pressures will be strong.
Presentation by Anna Mills, licensed CC BY NC 4.0.
Restrictions on access to AI during writing. Even if these would fix the problem in education, they wouldn’t work around text in other settings.
Presentation by Anna Mills, licensed CC BY NC 4.0.
AI Text Classification
Detectors: How good are they?
Q: Can software check if text is AI-generated?
A: There is software designed to classify text as “likely AI” or not.�
Presentation by Anna Mills, licensed CC BY NC 4.0.
First, a caution: ChatGPT doesn’t “know” if it wrote something!
Don’t paste student work into ChatGPT and ask, “Did you write this?”
Presentation by Anna Mills, licensed CC BY NC 4.0.
How does AI detection compare to (traditional) plagiarism detection?
Differences
Similarities
Presentation by Anna Mills, licensed CC BY NC 4.0.
100% reliable AI detection? Not likely
“We’re extremely unlikely to ever get a tool that can spot AI-generated text with 100% certainty. It’s really hard to detect AI-generated text because the whole point of AI language models is to generate fluent and human-seeming text….”
– “Why detecting AI-generated text is so difficult (and what to do about it)” by Melissa Heikkilä, MIT Technology Review, February 7, 2023
Presentation by Anna Mills, licensed CC BY NC 4.0.
AI Text Detection Software
Presentation by Anna Mills, licensed CC BY NC 4.0.
Turnitin claims a “less than 1/100 false positive rate,” but is that accurate?
Presentation by Anna Mills, licensed CC BY NC 4.0.
The Washington Post found an example of a false positive
GPTZero labeled the Bill of Rights “likely AI”
On April 17, 2023, I retested a popular Reddit experiment using GPTZero.me.
Presentation by Anna Mills, licensed CC BY NC 4.0.
Caution: false positives = false accusations?
Presentation by Anna Mills, licensed CC BY NC 4.0.
How will students feel when we tell them their writing may be wrongly flagged as likely AI?
To help students prepare for the possibility of false positives, Turnitin advises “Establish your voice: Make sure that your writing style and voice are evident to your instructor.”�
Presentation by Anna Mills, licensed CC BY NC 4.0.
Caution: Is detection a good use of money?
BUT
Presentation by Anna Mills, licensed CC BY NC 4.0.
Even according to Turnitin, detectors will often be inconclusive
Presentation by Anna Mills, licensed CC BY NC 4.0.
It’s currently simple for students to get around detection with paraphrase software like Quillbot.�
From the Washington Post on Turnitin’s AI detector: “It couldn’t spot the ChatGPT in papers we ran through Quillbot, a paraphrasing program that remixes sentences.”
Presentation by Anna Mills, licensed CC BY NC 4.0.
Free software explicitly designed to get around AI detectors
Presentation by Anna Mills, licensed CC BY NC 4.0.
Still, I’m not convinced we should give up on detection. Others disagree.
Presentation by Anna Mills, licensed CC BY NC 4.0.
I’m thinking of the situations in which we want to know if text is human or AI. I’m thinking of the limits on trust, restricting AI, and guessing, in education and beyond.
To me, the parallel is cybersecurity. We know it will never be reliable. We still find it helpful. It’s a tough problem worth working on.
AI text classification may be like this. We’ve only just begun to wrestle with it.
Presentation by Anna Mills, licensed CC BY NC 4.0.
Researchers see several ways forward for improving AI text classification
Presentation by Anna Mills, licensed CC BY NC 4.0.
Researchers from UMass Amherst and Google suggest that language models could keep a record of what they generate. They envision an API that “stores every sequence generated by their LLM in a database” and “offers an interface that allows users to enter candidate AI-generated text as a query.” A language model is used to check paraphrased text against the database of outputs.
Presentation by Anna Mills, licensed CC BY NC 4.0.
Turnitin is developing paraphrase detection
From Turnitin’s FAQ:
“Turnitin has been working on building paraphrase detection capabilities – ability to detect when students have paraphrased content either with the help of paraphrasing tools or re-written it themselves…We have plans for a beta release in 2023…for an additional cost.”
Presentation by Anna Mills, licensed CC BY NC 4.0.
Some students want teachers and students to use detectors. They don’t want to be at a competitive disadvantage.
Boston University students in a data science class with Professor Wesley Wildman developed a policy that allows different options for AI use but involves detectors extensively.
https://www.bu.edu/articles/2023/student-designed-policy-on-use-of-generative-ai-adopted-by-data-sciences-faculty/
Presentation by Anna Mills, licensed CC BY NC 4.0.
From the student-
developed policy adopted by the Data Sciences Department at Boston University
Presentation by Anna Mills, licensed CC BY NC 4.0.
Could student desire for detectors just be evidence of the pressures on them/identification with oppressive power structures they’ve learned?
My response: Perhaps, but we should still listen and respond. It seems condescending to dismiss a desire for classification software as false consciousness.
Presentation by Anna Mills, licensed CC BY NC 4.0.
Detection as deterrent: �Point out that what’s generated by AI might be labeled as AI, sooner or later
Presentation by Anna Mills, licensed CC BY NC 4.0.
What if everyone, including students, had access to classification software, free?
Presentation by Anna Mills, licensed CC BY NC 4.0.
Advocate for government support for AI text classification as a social good
If transparency is a social good in education and in other areas, let’s advocate for a government role in promoting human-nonhuman distinction to protect human communication
That’s my current position… What is yours?
While I don’t think we should use classification software punitively, I do think we should pursue it.
If used cautiously, it can help us maintain space for the human in textual communication.
Its existence can help promote
Presentation by Anna Mills, licensed CC BY NC 4.0.
Presentation by Anna Mills, licensed CC BY NC 4.0.
Articles about AI text detection especially in higher ed
What are your thoughts on whether and how we might move towards transparency around AI text?��Comments? Questions?
�Anna Mills�armills@marin.edu, @EnglishOER
Slides: https://bit.ly/TowardsTransparencyGPT��This presentation is shared under a CC BY NC 4.0 license.