Developer(s) | Edward Tian Alex Cui Yazan Mimi [1] |
---|---|
Initial release | 3 January 2023 |
Written in | Python |
Platform | Cloud computing |
Website | gptzero |
GPTZero is an artificial intelligence detection software developed to identify artificially generated text, such as those produced by large language models. [3] [4] [5] [6]
While GPTZero was praised for its efforts to prevent academic dishonesty, many news outlets criticized the tool's false positive rate, which can be especially harmful in academic settings. [7]
GPTZero was developed by Edward Tian, a Princeton University undergraduate student, and launched online in January 2023 in response to concerns about AI-generated usage in academic plagiarism. [8] [2] GPTZero said in May 2023 it raised over 3.5 million dollars in seed funding. [9] [10]
In the first week of its release, the GPTZero experienced 30,000 uses, which led to a crash. It was supported by the web app company Streamlit, who allocated more server resources in response. [11] In July 2024, it had 4 million users, compared to 1 million one year earlier. [12]
In summer 2024, GPTZero raised $10 million in Series A round funding. [13]
In September 2024, GPTZero announced an authorship tracking software that enables "to compile and share data about their writing process such as their copy/paste history, the number of editors they had, and how long editing took", in an effort "to move away from an all-or-nothing paradigm around AI writing towards a more nuanced one." [13]
GPTZero uses qualities it terms perplexity and burstiness to attempt determining if a passage was written by a AI. [14] According to the company, perplexity is how random the text in the sentence is, and whether the way the sentence is constructed is unusual or "surprising" for the application. It relies on language models, and the more such models, the higher the chance that a person did not write the text. [15] In contrast, burstiness compares sentences with each other, determining their similarity. Human text is more discontinuous, meaning humans tend to write with more sentence variation than AI. [2]
The academic community has attempted using GPTZero to tackle concerns about AI-generated content for plagiarism. [16] [15] [14] Educational institutions, including Princeton University, have discussed the use of GPTZero to combat AI-generated content in academic settings, with mixed opinions. [2] [17] In October 2023, GPTZero had partnered with the American Federation of Teachers. [18]
By 2024, Tian reported that GPTZero also "received a lot of adoption with hiring managers, with recruiting [and] cover letter analysis." [13]
In a March 2023 paper named "Can AI-Generated Text be Reliably Detected?", [19] computer scientists Vinu Sankar Sadasivan, Aounon Kumar, Sriram Balasubramanian, Wenxiao Wang, and Soheil Feizi from the University of Maryland demonstrate empirically and theoretically that several AI-text detectors are not reliable in practical scenarios. [20] [21]
Tech website Futurism tested the tool, and said that while the "results are impressive", based on its error rate, teachers relying on the tool would end up "falsely accusing nearly 20 percent of [innocent students] of academic misconduct". [22]
The Washington Post noted in August 2023 that GPTZero suffers from false positives, emphasizing that "even a small 'false positive' error rate means that some students could be wrongly accused [of academic misconduct]". [7]
News website Ars Technica commented that humans can still write sentences in a highly regular way, leading to false positives. The writer, Benj Edwards, went on to state that the perplexity score only concerns itself with what is "surprising" for the AI, leading to instances where highly common texts, such as the US Constitution, are labeled as likely AI-generated. [23]
A chatbot is a software application or web interface designed to have textual or spoken conversations. Modern chatbots are typically online and use generative artificial intelligence systems that are capable of maintaining a conversation with a user in natural language and simulating the way a human would behave as a conversational partner. Such chatbots often use deep learning and natural language processing, but simpler chatbots have existed for decades.
Turnitin is an Internet-based similarity detection service run by the American company Turnitin, LLC, a subsidiary of Advance Publications.
Plagiarism detection or content similarity detection is the process of locating instances of plagiarism or copyright infringement within a work or document. The widespread use of computers and the advent of the Internet have made it easier to plagiarize the work of others.
Plagiarism is the representation of another person's language, thoughts, ideas, or expressions as one's own original work. Although precise definitions vary depending on the institution, in many countries and cultures plagiarism is considered a violation of academic integrity and journalistic ethics, as well as of social norms around learning, teaching, research, fairness, respect, and responsibility. As such, a person or entity that is determined to have committed plagiarism is often subject to various punishments or sanctions, such as suspension, expulsion from school or work, fines, imprisonment, and other penalties.
Quizlet is a multi-national American company that provides tools for studying and learning. Quizlet was founded in October 2005 by Andrew Sutherland, who at the time was a 15-year old student, and released to the public in January 2007. Quizlet's primary products include digital flash cards, matching games, practice electronic assessments, and live quizzes. In 2017, 1 in 2 high school students used Quizlet. As of December 2021, Quizlet has over 500 million user-generated flashcard sets and more than 60 million active users.
Grammarly is a writing assistant. It reviews the spelling, grammar, and tone of a piece of writing as well as identifying possible instances of plagiarism. It can also can suggest style and tonal recommendations to users and produce writing from prompts with its generative AI capabilities.
OpenAI is an American artificial intelligence (AI) research organization founded in December 2015 and headquartered in San Francisco, California. Its stated mission is to develop "safe and beneficial" artificial general intelligence (AGI), which it defines as "highly autonomous systems that outperform humans at most economically valuable work". As a leading organization in the ongoing AI boom, OpenAI is known for the GPT family of large language models, the DALL-E series of text-to-image models, and a text-to-video model named Sora. Its release of ChatGPT in November 2022 has been credited with catalyzing widespread interest in generative AI.
Generative Pre-trained Transformer 3 (GPT-3) is a large language model released by OpenAI in 2020.
DALL-E, DALL-E 2, and DALL-E 3 are text-to-image models developed by OpenAI using deep learning methodologies to generate digital images from natural language descriptions known as "prompts".
You.com is an AI assistant that began as a personalization-focused search engine. While still offering web search capabilities, You.com has evolved to prioritize a chat-first AI assistant.
Hive is an American artificial intelligence company offering machine learning models via APIs to enterprise customers. Hive uses around 700,000 gig workers to train data for its models through its Hive Work app. One of Hive's major offerings is to provide automated content moderation services.
ChatGPT is a generative artificial intelligence (AI) chatbot developed by OpenAI and launched in 2022. It is based on the GPT-4o large language model (LLM). ChatGPT can generate human-like conversational responses, and enables users to refine and steer a conversation towards a desired length, format, style, level of detail, and language. It is credited with accelerating the AI boom, which has led to ongoing rapid investment in and public attention to the field of artificial intelligence. Some observers have raised concern about the potential of ChatGPT and similar programs to displace human intelligence, enable plagiarism, or fuel misinformation.
In the field of artificial intelligence (AI), a hallucination or artificial hallucination is a response generated by AI that contains false or misleading information presented as fact. This term draws a loose analogy with human psychology, where hallucination typically involves false percepts. However, there is a key difference: AI hallucination is associated with erroneous responses rather than perceptual experiences.
Generative artificial intelligence is a subset of artificial intelligence that uses generative models to produce text, images, videos, or other forms of data. These models learn the underlying patterns and structures of their training data and use them to produce new data based on the input, which often comes in the form of natural language prompts.
Artificial intelligence detection software aims to determine whether some content was generated using artificial intelligence (AI).
Since the public release of ChatGPT by OpenAI in November 2022, the integration of chatbots in education has sparked considerable debate and exploration. Educators' opinions vary widely; while some are skeptical about the benefits of large language models, many see them as valuable tools.
QuillBot is a software developed in 2017 that uses artificial intelligence to rewrite and paraphrase text.
Perplexity AI is a conversational search engine that uses large language models (LLMs) to answer queries. Its developer, Perplexity AI, Inc., is based in San Francisco, California.
Undetectable AI is an artificial intelligence content detection and modification software designed to identify and alter artificially generated text, such as that produced by large language models.
Copyleaks is a plagiarism detection platform that uses artificial intelligence (AI) to identify similar and identical content across various formats.