Artificial intelligence in Wikimedia projects

Last updated

Artificial intelligence is used in Wikipedia and other Wikimedia projects for the purpose of developing those projects. [1] [2] Human and bot interaction in Wikimedia projects is routine and iterative. [3]

Contents

Using artificial intelligence for Wikimedia projects

Various projects seek to improve Wikipedia and Wikimedia projects by using artificial intelligence tools.

ORES

The Objective Revision Evaluation Service (ORES) project is an artificial intelligence service for grading the quality of Wikipedia edits. [4] [5] The Wikimedia Foundation presented the ORES project in November 2015. [6]

Detox

Detox was a project by Google, in collaboration with the Wikimedia Foundation, to research methods that could be used to address users posting unkind comments in Wikimedia community discussions. [7] Among other parts of the Detox project, the Wikimedia Foundation and Jigsaw collaborated to use artificial intelligence for basic research and to develop technical solutions[ example needed ] to address the problem. In October 2016 those organizations published "Ex Machina: Personal Attacks Seen at Scale" describing their findings. [8] [9] Various popular media outlets reported on the publication of this paper and described the social context of the research. [10] [11] [12]

Bias reduction

In August 2018, a company called Primer reported attempting to use artificial intelligence to create Wikipedia articles about women as a way to address gender bias on Wikipedia. [13] [14]

Generative language models

In 2022, the public release of ChatGPT inspired more experimentation with AI and writing Wikipedia articles. A debate was sparked about whether and to what extent such large language models are suitable for such purposes in light of their tendency to generate plausible-sounding misinformation, including fake references; to generate prose that is not encyclopedic in tone; and to reproduce biases. [15] [16] As of May 2023, a draft Wikipedia policy on ChatGPT and similar large language models (LLMs) recommended that users who are unfamiliar with LLMs should avoid using them due to the aforementioned risks, as well as the potential for libel or copyright infringement. [16]

Using Wikimedia projects for artificial intelligence

Content in Wikimedia projects is useful as a dataset in advancing artificial intelligence research and applications. For instance, in the development of the Google's Perspective API that identifies toxic comments in online forums, a dataset containing hundreds of thousands of Wikipedia talk page comments with human-labelled toxicity levels was used. [17]

A 2012 paper reported that more than 1000 academic articles, including those using artificial intelligence, examine Wikipedia, reuse information from Wikipedia, use technical extensions linked to Wikipedia, or research communication about Wikipedia. [18] A 2017 paper described Wikipedia as the mother lode for human-generated text available for machine learning. [19]

A 2016 research project called "One Hundred Year Study on Artificial Intelligence" named Wikipedia as a key early project for understanding the interplay between artificial intelligence applications and human engagement. [20]

Related Research Articles

<span class="mw-page-title-main">Chatbot</span> Program that simulates conversation

A chatbot is a software application or web interface that is designed to mimic human conversation through text or voice interactions. Modern chatbots are typically online and use generative artificial intelligence systems that are capable of maintaining a conversation with a user in natural language and simulating the way a human would behave as a conversational partner. Such chatbots often use deep learning and natural language processing, but simpler chatbots have existed for decades.

An artificial general intelligence (AGI) is a hypothetical type of intelligent agent which, if realized, could learn to accomplish any intellectual task that human beings or animals can perform. Alternatively, AGI has been defined as an autonomous system that surpasses human capabilities in the majority of economically valuable tasks.Creating AGI is a primary goal of some artificial intelligence research and of companies such as OpenAI, DeepMind, and Anthropic. AGI is a common topic in science fiction and futures studies.

In the history of artificial intelligence, an AI winter is a period of reduced funding and interest in artificial intelligence research. The field has experienced several hype cycles, followed by disappointment and criticism, followed by funding cuts, followed by renewed interest years or even decades later.

Anthropic PBC is an American artificial intelligence (AI) startup company, founded by former members of OpenAI. Anthropic develops general AI systems and large language models. It is a public-benefit corporation, and has been connected to the effective altruism movement.

<span class="mw-page-title-main">Progress in artificial intelligence</span> How AI-related technologies evolve

Progress in artificial intelligence (AI) refers to the advances, milestones, and breakthroughs that have been achieved in the field of artificial intelligence over time. AI is a multidisciplinary branch of computer science that aims to create machines and systems capable of performing tasks that typically require human intelligence. Artificial intelligence applications have been used in a wide range of fields including medical diagnosis, economic-financial applications, robot control, law, scientific discovery, video games, and toys. However, many AI applications are not perceived as AI: "A lot of cutting edge AI has filtered into general applications, often without being called AI because once something becomes useful enough and common enough it's not labeled AI anymore." "Many thousands of AI applications are deeply embedded in the infrastructure of every industry." In the late 1990s and early 21st century, AI technology became widely used as elements of larger systems, but the field was rarely credited for these successes at the time.

<span class="mw-page-title-main">Aaron Halfaker</span> American computer scientist

Aaron Halfaker is a principal applied scientist at Microsoft Research. He previously served as a research scientist at the Wikimedia Foundation until 2020.

<span class="mw-page-title-main">OpenAI</span> Artificial intelligence research organization

OpenAI is a U.S. based artificial intelligence (AI) research organization founded in December 2015, researching artificial intelligence with the goal of developing "safe and beneficial" artificial general intelligence, which it defines as "highly autonomous systems that outperform humans at most economically valuable work". As one of the leading organizations of the AI Spring, it has developed several large language models, advanced image generation models, and previously, released open-source models. Its release of ChatGPT has been credited with starting the artificial intelligence spring.

Generative Pre-trained Transformer 3 (GPT-3) is a large language model released by OpenAI in 2020. Like its predecessor, GPT-2, it is a decoder-only transformer model of deep neural network, which supersedes recurrence and convolution-based architectures with a technique known as "attention". This attention mechanism allows the model to selectively focus on segments of input text it predicts to be most relevant. It uses a 2048-tokens-long context, float16 (16-bit) precision, and a hitherto-unprecedented 175 billion parameters, requiring 350GB of storage space as each parameter takes 2 bytes of space, and has demonstrated strong "zero-shot" and "few-shot" learning abilities on many tasks.

<span class="mw-page-title-main">ChatGPT</span> AI chatbot developed by OpenAI

ChatGPT is a chatbot developed by OpenAI and launched on November 30, 2022. Based on a large language model, it enables users to refine and steer a conversation towards a desired length, format, style, level of detail, and language. Successive prompts and replies, known as prompt engineering, are considered at each conversation stage as a context.

<span class="mw-page-title-main">Hallucination (artificial intelligence)</span> Confident unjustified claim by AI

In the field of artificial intelligence (AI), a hallucination or artificial hallucination is a response generated by an AI which contains false or misleading information presented as fact.

Sparrow is a chatbot developed by the artificial intelligence research lab DeepMind, a subsidiary of Alphabet Inc. It is designed to answer users' questions correctly, while reducing the risk of unsafe and inappropriate answers. One motivation behind Sparrow is to address the problem of language models producing incorrect, biased or potentially harmful outputs. Sparrow is trained using human judgements, in order to be more “Helpful, Correct and Harmless” compared to baseline pre-trained language models. The development of Sparrow involved asking paid study participants to interact with Sparrow, and collecting their preferences to train a model of how useful an answer is.

<span class="mw-page-title-main">Generative pre-trained transformer</span> Type of large language model

Generative pre-trained transformers (GPT) are a type of large language model (LLM) and a prominent framework for generative artificial intelligence. They are artificial neural networks that are used in natural language processing tasks. GPTs are based on the transformer architecture, pre-trained on large data sets of unlabelled text, and able to generate novel human-like content. As of 2023, most LLMs have these characteristics and are sometimes referred to broadly as GPTs.

The dead Internet theory is an online conspiracy theory that asserts that the Internet now consists mainly of bot activity and automatically generated content that is manipulated by algorithmic curation, marginalizing organic human activity. Proponents of the theory believe these bots are created intentionally to help manipulate algorithms and boost search results in order to ultimately manipulate consumers. Furthermore, some proponents of the theory accuse government agencies of using bots to manipulate public perception, stating: "The U.S. government is engaging in an artificial intelligence powered gaslighting of the entire world population." The date given for this "death" is generally around 2016 or 2017.

A large language model (LLM) is a language model notable for its ability to achieve general-purpose language generation. LLMs acquire these abilities by learning statistical relationships from text documents during a computationally intensive self-supervised and semi-supervised training process. LLMs are artificial neural networks, the largest and most capable of which are built with a transformer-based architecture. Some recent implementations are based on other architectures, such as recurrent neural network variants and Mamba.

<span class="mw-page-title-main">Bard (chatbot)</span> AI chatbot developed by Google

Bard is a conversational generative artificial intelligence chatbot developed by Google. Initially based on the LaMDA family of large language models (LLMs), it was later upgraded to PaLM and then to Gemini. Bard was developed as a direct response to the meteoric rise of OpenAI's ChatGPT, and was released in a limited capacity in March 2023 to lukewarm responses before expanding to other countries in May.

Poe is a service developed by Quora and launched in December 2022. It allows users to ask questions and obtain answers from a range of AI bots built on top of large language models (LLMs), including those from ChatGPT developer OpenAI, and other companies like Anthropic. It also has a subscription which allows users unlimited use of lightweight chatbot applications such as GPT-3.5, and provides access, with certain limitations, to more advanced artificial intelligence models such as GPT-4 and Claude 2.

GPTZero is an artificial intelligence detection software developed to identify artificially generated text, such as that produced by large language models.

Artificial intelligence detection software aims to determine whether some content was generated using artificial intelligence (AI).

Comparison of user features of chatbots refers to a comparison of the general user features of major chatbot applications or web interfaces, in a narrative format. It is a comparison of basic roles and the most prominent features. It does not encompass a full exhaustive comparison or description of all technical details of all chatbots. It also includes the most important features of the chatbots origins, historical development, and role.

References

  1. Marr, Bernard (17 August 2018). "The Amazing Ways How Wikipedia Uses Artificial Intelligence". Forbes.
  2. Gertner, Jon (18 July 2023). "Wikipedia's Moment of Truth - Can the online encyclopedia help teach A.I. chatbots to get their facts right — without destroying itself in the process? + comment". The New York Times . Archived from the original on 18 July 2023. Retrieved 19 July 2023.{{cite news}}: CS1 maint: bot: original URL status unknown (link)
  3. Piscopo, Alessandro (1 October 2018). "Wikidata: A New Paradigm of Human-Bot Collaboration?". arXiv: 1810.00931 [cs.HC].
  4. Simonite, Tom (1 December 2015). "Software That Can Spot Rookie Mistakes Could Make Wikipedia More Welcoming". MIT Technology Review.
  5. Metz, Cade (1 December 2015). "Wikipedia Deploys AI to Expand Its Ranks of Human Editors". Wired.
  6. Halfaker, Aaron; Taraborelli, Dario (30 November 2015). "Artificial intelligence service "ORES" gives Wikipedians X-ray specs to see through bad edits". Wikimedia Foundation.
  7. Research:Detox - Meta.
  8. Wulczyn, Ellery; Thain, Nithum; Dixon, Lucas (2017). "Ex Machina: Personal Attacks Seen at Scale". Proceedings of the 26th International Conference on World Wide Web. pp. 1391–1399. arXiv: 1610.08914 . doi:10.1145/3038912.3052591. ISBN   9781450349130. S2CID   6060248.
  9. Jigsaw (7 February 2017). "Algorithms And Insults: Scaling Up Our Understanding Of Harassment On Wikipedia". Medium.
  10. Wakabayashi, Daisuke (23 February 2017). "Google Cousin Develops Technology to Flag Toxic Online Comments". The New York Times.
  11. Smellie, Sarah (17 February 2017). "Inside Wikipedia's Attempt to Use Artificial Intelligence to Combat Harassment". Motherboard. Vice Media.
  12. Gershgorn, Dave (27 February 2017). "Alphabet's hate-fighting AI doesn't understand hate yet". Quartz.
  13. Simonite, Tom (3 August 2018). "Using Artificial Intelligence to Fix Wikipedia's Gender Problem". Wired.
  14. Verger, Rob (7 August 2018). "Artificial intelligence can now help write Wikipedia pages for overlooked scientists". Popular Science.
  15. Harrison, Stephen (2023-01-12). "Should ChatGPT Be Used to Write Wikipedia Articles?". Slate Magazine. Retrieved 2023-01-13.
  16. 1 2 Woodcock, Claire (2 May 2023). "AI Is Tearing Wikipedia Apart". Vice.
  17. "Google's comment-ranking system will be a hit with the alt-right". Engadget. 2017-09-01.
  18. Nielsen, Finn Årup (2012). "Wikipedia Research and Tools: Review and Comments". SSRN Working Paper Series. doi:10.2139/ssrn.2129874. ISSN   1556-5068.
  19. Mehdi, Mohamad; Okoli, Chitu; Mesgari, Mostafa; Nielsen, Finn Årup; Lanamäki, Arto (March 2017). "Excavating the mother lode of human-generated text: A systematic review of research that uses the wikipedia corpus". Information Processing & Management. 53 (2): 505–529. doi:10.1016/j.ipm.2016.07.003. S2CID   217265814.
  20. "AI Research Trends - One Hundred Year Study on Artificial Intelligence (AI100)". ai100.stanford.edu.

See also