Copyleaks thought everything in this article was written by AI, even though only a few paragraphs were.
Screenshot by Jon Reed/CNET
I tried again, this time asking Google's Gemini to add some copy to my existing story. Copyleaks again identified that 67.2% of the text matched what was online, but it also reported that 100% of the text may have been AI-generated. Even text I wrote was flagged, with some phrases, like "generative AI model," described as occurring more frequently in AI-written text.
Example: Totally AI-written
In a test of generative AI's ability to create things that are totally out of touch with reality, I asked it to write a news story as if the Cincinnati Bengals had won the Super Bowl. (In this fictional universe, Cincinnati beat the San Francisco 49ers by a score of 31-17.) When I ran the fake story through Copyleaks, it successfully identified it as entirely AI-written.
Enlarge Image
Copyleaks' AI Logic quickly realized this story about the Cincinnati Bengals winning the Super Bowl was written by an AI chatbot.
Screenshot by Jon Reed/CNET
What Copyleaks didn't do, however, is explain why. It said no results were found in its AI Source Match or its AI Phrases, but with a note: "There is no specific phrase that indicates AI. However, other criteria suggest that this text was generated by AI."
I tried again, this time with a different ChatGPT-generated story about the Bengals winning the Super Bowl 27-24 over the 49ers, and Copyleaks provided a more detailed explanation. It calculated the content was 98.7% AI-created, with a handful of phrases singled out. These included some seemingly innocent terms like "made several critical" and "testament to years of." It also included some strings of words that spread across multiple phrases or sentences, like "continues to evolve, the Bengals' future," which apparently occurred 317 times more frequently in the database's AI-generated content than in human text documents. (After raising the issue with the first attempt with Copyleaks, I tried it again and got similar results to this second test.)
Just to be sure it wasn't operating entirely on the fact that the Bengals have never won a Super Bowl, I asked ChatGPT to write an article about the Los Angeles Dodgers winning the World Series. Copyleaks found that 50.5% matched existing text online, but also reported it was 100% AI-written.
A high-profile example
Copyleaks did some testing of its own, using a recent example of a controversial alleged use of AI. In May, the news outlet NOTUS said that a report from the Trump administration's Make America Healthy Again Commission contained references to academic studies that did not exist. Researchers who were cited in the MAHA report told media outlets that they did not produce that work. Citations to nonexistent sources are a common result of AI hallucination, which is why it's important to check anything an LLM cites. The Trump administration defended the report, with a spokesperson blaming "minor citation and formatting errors" and stating that the substance of the report remains unchanged.
Copyleaks ran the report through its system, which reported finding 20.8% potential AI-written content. It found some sections around children's mental health raised red flags in its AI Phrases database. Some phrases that occurred far more often in AI-written text included "impacts of social media on their" and "The Negative Impact of Social Media on Their Mental Health."
Can an AI really detect AI-written text?
In my experience, the increased transparency from Copyleaks into how the tool works is a step forward for the world of AI detection, but this is still far from foolproof. There's still a troubling risk of false positives. In my testing, sometimes words I had written just hours before (and I know AI didn't play a role in them) could be flagged because of some of the phrasing. Still, Copyleaks was able to spot a bogus news article about a team that has never won a championship doing so.
Yamin said the goal isn't necessarily to be the ultimate source of truth but to provide people who need to assess whether and how AI has been used with tools to make better decisions. A human needs to be in the loop, but tools like Copyleaks can help with trust.
"The idea in the end is to help humans in the process of evaluating content," he said. "I think we're in an age where content is everywhere, and it's being produced more and more and faster than ever before. It's getting harder to identify content that you can trust."
Here's my take: When using an AI detector, one way to have more confidence is to look specifically at what is being flagged as possibly AI-written. The occasional suspicious phrase may be, and likely is, innocent. After all, there are only so many different ways you can rearrange words -- a compact phrase like "generative AI model" is pretty handy for us humans, same as for AI. But if it's several whole paragraphs? That may be more troubling.
AI detectors, just like that rumor that the em dash is an AI tell, can have false positives. A tool that is still largely a black box will make mistakes, and that can be devastating for someone whose genuine writing was flagged through no fault of their own.
I asked Yamin how human writers can make sure their work isn't caught in that trap. "Just do your thing," he said. "Make sure you have your human touch."