Skip to main content

Here’s how ChatGPT could solve its major plagiarism problem

ChatGPT is a wonderful tool but there’s a dark side to this advanced AI service that can write like an expert on almost any topic — plagiarism. When students that are supposed to be demonstrating their knowledge and understanding of a topic cheat by secretly using ChatGPT, it invalidates testing and grading. AI skills are great but aren’t the only subject that students should learn.

Policing this problem has proven to be difficult. Since ChatGPT has been trained on a vast dataset of human writing, it’s nearly impossible for an instructor to identify whether an essay was created by a student or a machine. Several tools have been created that attempt to recognize AI-generated writing, but the accuracy was too low to be useful.

Amidst rising concerns from educators and bans on students using ChatGPT, Business Insider reports that OpenAI is working on a solution to this problem. A recent tweet from Tom Goldstein, Associate Professor of machine learning at the University of Maryland, explained how accurate it might be at detecting watermarked text that’s written by ChatGPT.

#OpenAI is planning to stop #ChatGPT users from making social media bots and cheating on homework by "watermarking" outputs. How well could this really work? Here's just 23 words from a 1.3B parameter watermarked LLM. We detected it with 99.999999999994% confidence. Here's how 🧵 pic.twitter.com/pVC9M3qPyQ

— Tom Goldstein (@tomgoldsteincs) January 25, 2023

Any tool that could identify plagiarism with nearly 100% accuracy would settle this discussion quickly and alleviate any concerns. According to Goldstein, one solution is to make the large language model (LLM) pick from a limited vocabulary of words, forming a whitelist that is okay for the AI to use and a blacklist of words that are forbidden. If an unnaturally large number of whitelist words show up in a sample, that would suggest it was generated by the AI.

This simplistic approach would be too restrictive since it’s hard to predict which words might be necessary for a discussion when working one word at a time, as most LLMs do. Goldstein suggests that ChatGPT could be given the ability to look ahead further than one word so it can plan a sentence that can be filled with whitelisted words while still making sense.

ChatGPT made a big splash when it entered the community writing pool and can be a great teaching aide as well. It’s important to introduce artificial intelligence in schools since it will clearly be an important technology to understand in the future, but it will continue to be controversial until the issue of plagiarism is addressed.

Editors' Recommendations

Alan Truly
Computing Writer
Alan is a Computing Writer living in Nova Scotia, Canada. A tech-enthusiast since his youth, Alan stays current on what is…
This one image breaks ChatGPT each and every time
A laptop screen shows the home page for ChatGPT, OpenAI's artificial intelligence chatbot.

Sending images as prompts to ChatGPT is still a fairly new feature, but in my own testing, it works fine most of the time. However, someone's just found an image that ChatGPT can't seem to handle, and it's definitely not what you expect.

The image, spotted by brandon_xyzw on X (formerly Twitter), presents some digital noise. It's nothing special, really -- just a black background with some vertical lines all over it. But if you try to show it to ChatGPT, the image breaks the chatbot each and every time, without fail.

Read more
Researchers just unlocked ChatGPT
ChatGPT versus Google on smartphones.

Researchers have discovered that it is possible to bypass the mechanism engrained in AI chatbots to make them able to respond to queries on banned or sensitive topics by using a different AI chatbot as a part of the training process.

A computer scientists team from Nanyang Technological University (NTU) of Singapore is unofficially calling the method a "jailbreak" but is more officially a "Masterkey" process. This system uses chatbots, including ChatGPT, Google Bard, and Microsoft Bing Chat, against one another in a two-part training method that allows two chatbots to learn each other's models and divert any commands against banned topics.

Read more
OpenAI and Microsoft sued by NY Times for copyright infringement
A phone with the OpenAI logo in front of a large Microsoft logo.

The New York Times has become the first major media organization to take on AI firms in the courts, accusing OpenAI and its backer, Microsoft, of infringing its copyright by using its content to train AI-powered products such as OpenAI's ChatGPT.

In a lawsuit filed in Federal District Court in Manhattan, the media giant claims that “millions” of its copyrighted articles were used to train its AI technologies, enabling it to compete with the New York Times as a content provider.

Read more