Why I'm using AI detection after all, alongside many other strategies

Why I'm using AI detection after all, alongside many other strategies

I argued against use of AI detection in college classrooms for two years, but my perspective has shifted. I ran into the limits of my current approaches last semester, when a first-year writing student persisted in submitting work that was clearly not his own, presenting document history that showed him typing the work (maybe he. typed it and maybe he used an autotyper). He only admitted to the AI use and apologized for wasting my time when he realized that I was not going to give him credit and that if he initiated an appeals process, the college would run his writing through detection software. I liked this student, had met with him previously and encouraged him to build confidence in his own voice and helped him find a research topic that interested him. I didn't think he would be well served by an AI on an autopaper, so I was glad that AI detection existed.

I'm also influenced by recent research that suggests detection is likely not biased against English language learners after all, that educators are not as good as we think we are at distinguishing AI from student writing on our own, and that some detection systems are pretty accurate when it comes to naive copy/paste AI use. Christopher Ostro's slide deck lit review of recent research on detection has been invaluable. He discusses this research in an engaging episode of Bonni Stachowiak's wonderful podcast Teaching in Higher Ed. Dr. Tricia Bertram Gallant 's leadership on academic integrity and AI has influenced me over the last two years as well. She has shown levelheaded willingness to consider a possible role for detection even as she promotes other approaches as more important and effective.

As I teach composition online asychronously this semester, I'm incorporating Turnitin alongside process tracking, writing process assignments, social annotation, lots of student choice, peer review and tutoring, video assignments, and clear messages about the purpose of each activity and the value of the writing process. I love that Phillip Dawson has described this kind of laying of strategies as a "swiss cheese" approach and others have used the mosaic metaphor (I'm having trouble tracking down who). I've described my combination of approaches in slides for a recent presentation on Academic Integrity and AI.

What about the risk that AI detection will lead to false accusations? It's real, and I let students know I'm aware that detection yields some false positives. I will never trust AI detection as firm evidence, and I am not punishing students. If a student discusses an essay with me, shows process history, and denies AI use, I will give them the benefit of the doubt even if the detector says "AI." If they aren't able to discuss their writing, I ask students to rewrite.

Christopher Ostro put it in a way that resonates for me: "I think AI detection has a place, but its place is limited." In most human endeavors, some accountability structures are important even when we design for intrinsic motivation. And we don't have perfect options here. The options are fewer in online asynchronous classes that allow many working-class students and parents to access college.

I know so many colleagues I respect are highly critical of AI detection, seeing it as signaling antagonism toward students. Christopher Ostro clarifies that his purpose is not to punish students but to provide some accountability that, in the end, encourages learning and shows that we care. He says, "I am not a cop teacher. I am not someone who likes catching cheaters. I’m not someone who wants that to be a big part of my job. Honestly, it’s the least fun part of teaching, but it’s also it is still a part of the job."

I'm a member of the MLA/CCCC Joint Task Force on Writing and AI, a group that has put out strong cautions about it in our working paper on Generative AI and Policy Development. There, we argue that "Tools for detection and authorship verification in GAI use should be used with caution and discernment or not at all." We write, "For those who decide to use AI detectors, please consider the following questions: What steps have you taken to substantiate a positive detection? What other kinds of engagement with the student’s writing affirms your decision to assign a failing grade outside the AI detector’s claim that the text was AI generated?" We also emphasize that "any technological approaches to academic integrity should respect legal, privacy, nondiscrimination, and data rights of students."

I have tried to use detection and process tracking in ways that I hope address those concerns. I invite students to comment frankly on syllabus policies. If students don't want to share process history or if they object to AI detection, I invite them to meet with me to chat about the essay instead. My approach is a work in progress in a changing landscape. Next up: an anonymous survey to see what more I can find out about what students think.

Sherry Wynn Perdue

Director, Oakland University Writing Center at Oakland University

9 小时前

Thank you for sharing the process you made to rethink your practice in light of new evidence. This is the best of what we hope to teach our students. Everyday we make decisions informed by incomplete information , decisions that are t simple matters of what is true or false or good or evil. Either that in mind, we should be willing to change our practices in the face of compelling new information.

Anne Murphy

We Do Enterprise AI Adoption | Founder, #SheLeadsAI | Fundraising Consultant and Coach | Mom | Professional Speaker

5 天前

How long do you plan to use this as your approach to AI in coursework? If there is a case to use detectors, it's super temporary. Wouldn't it be sooooo much more productive to throw the limited time, energy, and creativity into changing the way the system teaches and evaluates? I'm stumped as to why people are still leaning on this band-aid.

回复
Joel Gladd

Department Chair of Integrated Studies at College of Western Idaho

1 周

I loved Chris’s overview as well, super helpful. There are so many over-generalizations about this issue and I like that you’re modeling how to remain tactical and focused.

Dr. Shannon H. Doak ??

Edtech and Innovation Leader. Speaker, Author, Lucky Father and Husband, #AI Enthusiast, #PoeCreator #CoffeeLover, and #HomeBarista #Bahai | Currently, I am the Director of Technology at Nanjing International School

1 周

I am concerned that we will not change our assessment practice if we use AI Detectors. While I am all for academic integrity and honesty, and also feel that AI should only be used on areas of the curriculum that are not being assessed. I however, think that relying on AI detectors may make it harder to move forward. I will dive deeper into your post and read more later....these are only my initial thoughts. Maybe after reading my thoughts will shift as well.

Jennifer Stevens

Director, Instructional Technology Group at Emerson College

2 周

I highly appreciate your nuance! We've been talking about this in our institution, and I've been on the no side of AI detection for three reasons: adversarial relationship, cost, and accuracy. I agree that the most important part of this is preserving the student/teacher relationship, and right now we're seeing students readily confess to AI use without having to resort to detection. I also agree in theory that AI detection could be useful as part of a Swiss cheese approach! The cost is so prohibitive, though, and what you get in return doesn't seem accurate enough to justify it. I've been looking at this meta-analysis (https://journals.sfu.ca/jalt/index.php/jalt/article/view/1369)and was discouraged by their conclusion: "Lastly, one major aspect flagged by the main findings of the 17 reviewed articles is the inconsistency of the detection efficacy of all the tested AI detectors and all the tested anti-plagiarism detection tools. To this end, both sets of AI detection tools lacked detection reliability."

要查看或添加评论,请登录

Anna Mills的更多文章

其他会员也浏览了