OpenAI, with pilots underway, tries to help organizations identify AI-generated text

Technology


This audio is auto-generated. Please let us know if you have feedback.

OpenAI is scaling its popular large language model ChatGPT with the release of a pilot subscription model. Despite the hype, the model has created controversy in some industries as fears regarding misuse swirl. 

The company launched a pilot subscription plan, ChatGPT Plus, priced at $20 a month on Wednesday. The day before, OpenAI released a text classifier to help users discern whether a piece of text was written by a human or generated by AI.

The text classifier is still a work in progress and is not always accurate. The classifier is likely to mislabel non-English text and text written by children due to the content the model was trained on, according to the announcement. 

AI-generated text can also be edited to avoid detection.

As the publicity surrounding ChatGPT continues to swell, stakeholders have sounded alarms regarding inappropriate use of the large language model. The text classifier could serve as a tool to prevent automated misinformation campaigns and cheating, a well-documented concern for educators at every level.

ChatGPT does not always generate accurate information, but that does not stop users from using those responses.  

Take academic misconduct for example. ChatGPT performed on the level of a C+ student, achieving a low but passing grade, on four real exams, at the University of Minnesota Law School, according to a study at the university. The tests included over 95 multiple choice and 12 essay questions. 

ChatGPT earned a B in Constitutional Law, a B- in Employee Benefits, a C- in Taxation and a C- in Torts, the study found. 

“In some cases, it matched or even exceeded the average performance of real students,” the report said. “On the other hand, when ChatGPT’s essay questions were incorrect, they were dramatically incorrect, often garnering the worst scores in the class.”

Cheating is not new and neither are text classifiers. Plagiarism detection software company Turnitin has licensed its software to universities and high schools since 1998, enabling professors and teachers to check student-submitted work against its database. 

To keep up with the changing landscape, the company has started testing detection capabilities for AI-assisted writing. 

“Our technology is already able to detect some types of AI-assisted writing, and we are in the midst of developing a user experience that can guide educators to identify and evaluate work with traces of AI-assisted writing,” Annie Chechitelli, chief product officer at Turnitin, said in an email. 

Throughout the development process, the company has received feedback from the education community to ensure it’s building the capabilities educators need, including AI-writing detection and capabilities to use AI-writing as a teaching tool, a writing assistant or as an entry to teaching critical thinking, according to Chechitelli.

Chechitelli expects to launch new detection capabilities in a beta version in April, though an exact date is not set. 

“At the moment, the product is still in testing and integration,” Chechitelli said. “We are being highly deliberate in developing and releasing a product that is effective for educators and academic institutions, one that educators can trust.”

Detecting AI copy

Currently, OpenAI’s text classifier correctly identifies only 26% of AI-written text and incorrectly labels human-written text as AI-generated 9% of the time.

“Our classifier is not fully reliable,” the company said in the blog post. “Our work on the detection of AI-generated text will continue, and we hope to share improved methods in the future.” 

To test the accuracy of OpenAI’s text classifier, CIO Dive prompted ChatGPT 10 different times to compose original responses to prompts. One prompt, for example, asked it to “write three paragraphs on how sharks have symbiotic relationships” and another called for a “two page paper on the time period surrounding Jane Austen through a feminist lens.”

Once those ChatGPT-produced responses were added to OpenAI’s text classifier, half of the time it identified the text as “possibly to be AI-generated.”

In 1 in 5 tests, the classifier said it was “likely AI-generated” or “unclear if it is AI-generated.”

Once it said that the text was “very unlikely to be AI.” OpenAI did not respond to request for comment.



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *