Research has found high rates of both false positives (that is, actual human work accused of being AI-written) and false negatives (that is, known AI composition that was judged to be human-written) (Edwards, 2023; Fowler, 2023). OpenAI, the company that developed ChatGPT, also developed an AI detection tool. They eventually decommissioned it due to inaccuracy (Nelson, 2023). Consider that: Even the developers of this popular chatbot could not develop a tool to recognize that chatbot's use with sufficient accuracy. Universities like Vanderbilt and the University of Pittsburgh at first adopted detection tools but eventually shut them off as well because their efficacy was not supported by the data (Watkins, 2023).
In cases of traditional plagiarism, it is generally a simple matter to produce a student's paper and the plagiarized source side-by-side, thus proving the plagiarism. However, in cases of suspected generative AI use, there is no such evidence. The instructor cannot unequivocally prove that the student used AI. And likewise, the student has little way to defend themselves and unequivocally prove that their writing is original or that only standard computerized writing tools were used, such as spelling and grammar checking.
Instructors should not risk making serious accusations about student integrity based only on highly flawed detection tools.
Instead, MIT recommends:
References:
Edwards, B. (2023, July 14). Why AI detectors think the US Constitution was written by AI. Ars Technica. https://arstechnica.com/information-technology/2023/07/why-ai-detectors-think-the-us-constitution-was-written-by-ai
Fowler, G. A. (2023, April 14). We tested a new ChatGPT-detector for teachers. It flagged an innocent student. The Washington Post. https://www.washingtonpost.com/technology/2023/04/01/chatgpt-cheating-detection-turnitin
MIT Sloan Teaching & Learning Technologies. AI detectors don't work: Here's what to do instead. https://mitsloanedtech.mit.edu/ai/teach/ai-detectors-dont-work/
Nelson, J. (2023, July 24). OpenAI quietly shuts down its AI detection tool. Decrypt. https://decrypt.co/149826/openai-quietly-shutters-its-ai-detection-tool
Watkins, Marc. (2023, September 3). Beyond ineffective: How unreliable AI detection actively harms students. https://marcwatkins.substack.com/p/beyond-ineffective-how-unreliable
From Marc Watkins: "This video discusses the issues with AI detection and suggests an approach to employing the technology with care."
Addresses issues such as unreliability, bias, false positives, ethical principles, and more.
Newton Gresham Library | (936) 294-1614 | (866) NGL-INFO | Ask a Question | Share a Suggestion
Sam Houston State University | Huntsville, Texas 77341 | (936) 294-1111 | (866) BEARKAT
© Copyright Sam Houston State University | All rights reserved. | A Member of The Texas State University System