Partnership on AI

Last updated
Logo of the Partnership on AI Partnership on AI logo.svg
Logo of the Partnership on AI

Partnership on Artificial Intelligence to Benefit People and Society, otherwise known as Partnership on AI, is a nonprofit coalition committed to the responsible use of artificial intelligence. Coming into inception in September 2016, PAI (Partnership on AI) grouped together members from over 90 companies and non-profits in order to explore best practice recommendations for the tech community. [1]

Contents

History

The Partnership on AI was publicly announced on September 28, 2016 with founding members Amazon, Facebook, Google, DeepMind, Microsoft, and IBM, with interim co-chairs Eric Horvitz of Microsoft Research and Mustafa Suleyman of DeepMind. [2] [3] [4] [5] More than 100 partners from academia, civil society, industry, and nonprofits are member organizations in 2019. [6]

In January 2017, Apple head of advanced development for Siri, Tom Gruber, joined the Partnership on AI's board. [7] In October 2017, Terah Lyons joined the Partnership on AI as the organization's founding executive director. [8] Lyons brought to the organization her expertise in technology governance, with a specific focus in machine intelligence, AI, and robotics policy, having formerly served as Policy Advisor to the United States Chief Technology Officer Megan Smith. Lyons was succeeded by Partnership on AI board member Rebecca Finlay as interim executive director. Finlay was named CEO of Partnership on AI on October 26, 2021.

In October 2017, Terah Lyons joined the Partnership on AI as the organization's founding executive director. Lyons brought to the organization her expertise in technology governance, with a specific focus in machine intelligence, AI, and robotics policy, having formerly served as Policy Advisor to the United States Chief Technology Officer Megan Smith. Lyons was succeeded by Partnership on AI board member Rebecca Finlay as interim executive director. Finlay was named CEO of Partnership on AI on October 26, 2021.

In October 2018, Baidu became the first Chinese firm to join the Partnership. [9]

In November 2020 the Partnership on AI announced the AI Incident Database (AIID), [10] which is a tool to identify, assess, manage, and communicate AI risk and harm.

In August 2021, the Partnership on AI submitted a response to the National Institute of Standards and Technology (NIST). The response provided examples of PAI’s work related to AI risk management, such as the Safety Critical AI report on responsible publication of AI research, the ABOUT ML project on documentation and transparency in machine learning lifecycles, and the AI Incident Database. [11] The response also highlighted how the AI Incident Database involves some of the minimum attributes in NIST’s AI RMF, such as being consensus-driven, risk-based, adaptable, and consistent with other approaches to managing AI risk. [11]

On October 26, 2021, Rebecca Finlay was named CEO. [12]

In February 2023, the Partnership on AI (PAI) launched a novel framework aimed at guiding the ethical development and use of synthetic media. This initiative was backed by a variety of initial partners, including notable entities such as Adobe, BBC, CBC/Radio-Canada, Bumble, OpenAI, TikTok, WITNESS, and synthetic media startups Synthesia, D-ID, and Respeecher. The framework, which emphasizes transparency, creativity, and safety, was the result of a year-long collaborative process involving contributions from a wide range of stakeholders, including synthetic media startups, social media platforms, news organizations, advocacy groups, academic institutions, policy professionals, and public commenters. [13]

Mission and Principles

Partnership on AI has a multiple pronged approach to achieve impact. Their initiatives are separated into five different programs: AI and media integrity; AI, work, and the economy; justice, transparency, and accountability; inclusive research and design; and security for AI. These programs aim to produce value through specific outputs, methodological tools, and articles. [14]

Through the program on AI & Media Integrity, PAI actively endeavors to establish best practices that ensure AI's positive influence on the global information ecosystem. Recognizing the potential for AI to facilitate harmful online content and amplify existing negative narratives, PAI is committed to mitigating these risks and fostering a responsible AI presence. [15]

The AI, Labor, and the Economy program serves as a collaborative platform, uniting economists, worker representative organizations, and PAI's partners to formulate a cohesive response on how AI can contribute to an inclusive economic future. The recent release of PAI's "Guidelines for AI and Shared Prosperity" on June 7, 2023, outlines a blueprint for the judicious use of AI across various stages, guiding organizations, policymakers, and labor entities. [16]

The Fairness, Transparency, and Accountability program, in conjunction with the Inclusive Research & Design program, strives to reshape the AI landscape towards justice and fairness. By exploring the intersections between AI and fundamental human values, the former establishes guidelines for algorithmic equity, explainability, and responsibility. Simultaneously, the latter empowers communities by providing guidelines on co-creating AI solutions, fostering inclusivity throughout the research and design process. [17] [18]

The Safety Critical AI program addresses the growing deployment of AI systems in pivotal sectors like medicine, finance, transportation, and social media. With a focus on anticipating and mitigating potential risks, the program brings together partners and stakeholders to develop best practices that span the entire AI research and development lifecycle. Notable initiatives include the establishment of the AI incident Database, formulation of norms for responsible publication, and the creation of the innovative AI learning environment SafeLife. [19]

The association is also built of thematic foundations that drive Partnership on AI's focus. Atop the programs mentioned above, Partnership on AI looks to expand upon the social impact of AI, encouraging positive social utility. The organization has highlighted potential benefits of AI within public welfare, education, sustainability, etc. With these specific use cases, Partnership on AI is developing an ethical framework in which to analyze and AI's measure of ethical efficacy. The ethical framework places an emphasis on inclusive participatory practices that enhance equity in AI. [20]

Programs and initiatives

The Partnership on AI has been involved in several initiatives aimed at promoting the responsible use of AI. One of their key initiatives is the development of a framework for the safe deployment of AI models. This framework guides model providers in developing and deploying AI models in a manner that ensures safety for society and can adapt to evolving capabilities and uses. [21]

In collaboration with DeepMind, the Partnership on AI has also launched a study to investigate the high attrition rates among women and minoritized individuals in tech. [22]

Recognizing the importance of explainability in AI, the Partnership on AI hosted a one-day, in-person workshop focused on the deployment of “explainable artificial intelligence” (XAI). This event brought together experts from various industries to discuss and explore the concept of XAI. [23]

Terah Lyons, Executive Director, Partnership on AI speaking at the AI for Good Global Summit 2018 15-17 May 2018, Geneva AI for Good Global Summit 2018 (41408771354).jpg
Terah Lyons, Executive Director, Partnership on AI speaking at the AI for Good Global Summit 2018 15–17 May 2018, Geneva

In an effort to support information integrity, the Partnership on AI collaborated with First Draft to investigate effective strategies for addressing deceptive content online. [24] This initiative reflects the organization’s methodical approach to identifying and promoting best practices in AI.

The Partnership on AI is also creating resources to facilitate effective engagement between AI practitioners and impacted communities. [25]

In November 2020, the Partnership on AI announced the AI Incident Database (AIID), a project dedicated to indexing the collective history of harms or near harms realized in the real world by the deployment of artificial intelligence systems. The AIID, which shifted to a new special-purpose independent non-profit in 2022, serves as a valuable resource for understanding and mitigating the potential risks associated with AI. [26]

Most recently, PAI conducted the PAI's 2023 Policy Forum. This event, held in London, was a gathering of diverse stakeholders to explore recent trends in AI policy globally and strategies for ensuring AI safety. During the event, the Partnership on AI (PAI) unveiled their "Guidance for Safe Foundation Model Deployment" for public feedback. This guidance, shaped by the Safety Critical AI Steering Committee and contributions from PAI's worldwide network, offers flexible principles for managing risks linked to large-scale AI implementation. Participants included policymakers, AI professionals, philanthropy and civil society members, and academic experts. [27]

Partners and members

The Board of Directors of the Partnership on AI (PAI) as of 2023 includes:

Criticisms

In October 2020, Access Now, announced its official resignation from PAI in a letter. Access Now stated that it had found that there was an increasingly smaller role for civil society to play within PAI and that PAI had not influenced or changed the attitude of member companies or encouraged them to respond to or consult with civil society on a systematic basis. Access Now also expressed its disagreement with PAI’s approach to AI ethics and risk assessment, and its advocacy for an outright ban on technologies that are fundamentally incompatible with human rights, such as facial recognition or other biometric technologies that enable mass surveillance. [29]

Related Research Articles

<span class="mw-page-title-main">Risk management</span> Identification, evaluation and control of risks

Risk management is the identification, evaluation, and prioritization of risks followed by coordinated and economical application of resources to minimize, monitor, and control the probability or impact of unfortunate events or to maximize the realization of opportunities.

The ethics of artificial intelligence covers a broad range of topics within the field that are considered to have particular ethical stakes. This includes algorithmic biases, fairness, automated decision-making, accountability, privacy, and regulation. It also covers various emerging or potential future challenges such as machine ethics, lethal autonomous weapon systems, arms race dynamics, AI safety and alignment, technological unemployment, AI-enabled misinformation, how to treat certain AI systems if they have a moral status, artificial superintelligence and existential risks.

The United Nations Interregional Crime and Justice Research Institute (UNICRI) is one of the five United Nations Research and Training Institutes. The institute was founded in 1968 to assist the international community in formulating and implementing improved policies in the field of crime prevention and criminal justice. Its work currently focuses on Goal 16 of the 2030 Agenda for Sustainable Development, that is centred on promoting peaceful, just and inclusive societies, free from crime and violence.

Palantir Technologies Inc. is a public American company that specializes in software platforms for big data analytics. Headquartered in Denver, Colorado, it was founded by Peter Thiel, Nathan Gettings, Joe Lonsdale, Stephen Cohen, and Alex Karp in 2003. The company's name is derived from The Lord of the Rings where the magical palantíri were "seeing-stones," described as indestructible balls of crystal used for communication and to see events in other parts of the world.

<span class="mw-page-title-main">Eric Horvitz</span> American computer scientist, and Technical Fellow at Microsoft

Eric Joel Horvitz is an American computer scientist, and Technical Fellow at Microsoft, where he serves as the company's first Chief Scientific Officer. He was previously the director of Microsoft Research Labs, including research centers in Redmond, WA, Cambridge, MA, New York, NY, Montreal, Canada, Cambridge, UK, and Bangalore, India.

PhotoDNA is a proprietary image-identification and content filtering technology widely used by online service providers.

<span class="mw-page-title-main">Netwrix</span>

Netwrix is a Frisco, Texas-based private IT security software company that develops software to help companies identify and secure sensitive data and assist with compliance auditing. After eight acquisitions the company's team geographically expanded to Latin America, UK, Germany, France, Asia, USA as well as other countries. The company's flagship products are Netwrix Auditor and StealthAUDIT that help information security and governance professionals manage sensitive, regulated and business-critical data.

<span class="mw-page-title-main">Mustafa Suleyman</span> British entrepreneur and activist

Mustafa Suleyman is a British artificial intelligence (AI) entrepreneur. He is the CEO of Microsoft AI, and the co-founder and former head of applied AI at DeepMind, an AI company acquired by Google. After leaving DeepMind, he co-founded Inflection AI, a machine learning and generative AI company, in 2022.

<span class="mw-page-title-main">Future of Life Institute</span> International nonprofit research institute

The Future of Life Institute (FLI) is a nonprofit organization which aims to steer transformative technology towards benefiting life and away from large-scale risks, with a focus on existential risk from advanced artificial intelligence (AI). FLI's work includes grantmaking, educational outreach, and advocacy within the United Nations, United States government, and European Union institutions.

RLDatix is a global enterprise software company offering software and services tailored to healthcare organizations. The technology platform is designed to support hospitals and other providers with risk mitigation, regulatory compliance, and workforce management resources.

OpenAI is an American artificial intelligence (AI) research organization founded in December 2015 and headquartered in San Francisco, California. Its mission is to develop "safe and beneficial" artificial general intelligence (AGI), which it defines as "highly autonomous systems that outperform humans at most economically valuable work". As a leading organization in the ongoing AI boom, OpenAI is known for the GPT family of large language models, the DALL-E series of text-to-image models, and a text-to-video model named Sora. Its release of ChatGPT in November 2022 has been credited with catalyzing widespread interest in generative AI.

The AI Now Institute is an American research institute studying the social implications of artificial intelligence and policy research that addresses the concentration of power in the tech industry. AI Now has partnered with organizations such as the Distributed AI Research Institute (DAIR), Data & Society, Ada Lovelace Institute, New York University Tandon School of Engineering, New York University Center for Data Science, Partnership on AI, and the ACLU. AI Now has produced annual reports that examine the social implications of artificial intelligence. In 2021-2, AI Now’s leadership served as a Senior Advisors on AI to Chair Lina Khan at the Federal Trade Commission. Its executive director is Amba Kak.

<span class="mw-page-title-main">Terah Lyons</span> American technology policy scholar

Terah Lyons is known for her work in the field of artificial intelligence and technology policy. Lyons was the executive director of the Partnership on AI and was a policy advisor to the United States Chief Technology Officer Megan Smith in President Barack Obama's Office of Science and Technology Policy.

ACM Conference on Fairness, Accountability, and Transparency is a peer-reviewed academic conference series about ethics and computing systems. Sponsored by the Association for Computing Machinery, this conference focuses on issues such as algorithmic transparency, fairness in machine learning, bias, and ethics from a multi-disciplinary perspective. The conference community includes computer scientists, statisticians, social scientists, scholars of law, and others.

Artificial Intelligence in India refers to the development, adoption, and impact of artificial intelligence (AI) technologies in the country. The AI market in India is projected to reach $8 billion by 2025, growing at a compound annual growth rate (CAGR) of over 40% from 2020 to 2025.

Regulation of artificial intelligence is the development of public sector policies and laws for promoting and regulating artificial intelligence (AI). It is part of the broader regulation of algorithms. The regulatory and policy landscape for AI is an emerging issue in jurisdictions worldwide, including for international organizations without direct enforcement power like the IEEE or the OECD.

<span class="mw-page-title-main">G42 (company)</span> Artificial Intelligence company

Group 42 Holding Ltd, doing business as G42, is an Emirati artificial intelligence (AI) development holding company based in Abu Dhabi, founded in 2018. The organization is focused on AI development across various industries including government, healthcare, finance, oil and gas, aviation, and hospitality. Tahnoun bin Zayed Al Nahyan, UAEs national security advisor is the controlling shareholder and chairs the company. Because G42 had strong ties to China, U.S. authorities have been concerned that G42 serves as a channel through which sophisticated U.S. technology is diverted to Chinese companies or the government. As of February 2024, G42 divested its stakes in China.

<span class="mw-page-title-main">Algorithmic Justice League</span> Digital advocacy non-profit organization

The Algorithmic Justice League (AJL) is a digital advocacy non-profit organization based in Cambridge, Massachusetts. Founded in 2016 by computer scientist Joy Buolamwini, the AJL uses research, artwork, and policy advocacy to increase societal awareness regarding the use of artificial intelligence (AI) in society and the harms and biases that AI can pose to society. The AJL has engaged in a variety of open online seminars, media appearances, and tech advocacy initiatives to communicate information about bias in AI systems and promote industry and government action to mitigate against the creation and deployment of biased AI systems. In 2021, Fast Company named AJL as one of the 10 most innovative AI companies in the world.

<span class="mw-page-title-main">Artificial Intelligence Act</span> 2024 European Union regulation on artificial intelligence

The Artificial Intelligence Act is a European Union regulation concerning artificial intelligence (AI). It establishes a common regulatory and legal framework for AI within the European Union (EU). It came into force on 1 August 2024, with provisions coming into operation gradually over the following 6 to 36 months.

<span class="mw-page-title-main">Elham Tabassi</span>

Elham Tabassi is an engineer and government leader. She was listed on the inaugural TIME100 Most Influential People in AI. Tabassi led the creation of the United States Artificial Intelligence Risk Management Framework, adopted by both industry and government. Tabassi was selected to serve on the National Artificial Intelligence (AI) Research Resource Task Force. Tabassi began her career in government at the National Institute of Standards and Technology, pioneering various machine learning and computer vision projects with applications in biometrics evaluation and standards, included in over twenty five publications. Her research has been deployed by the FBI and Department of Homeland Security.

References

  1. Belfield, Haydn (2020-02-07). "Activism by the AI Community: Analysing Recent Achievements and Future Prospects". Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society. AIES '20. New York, NY, USA: Association for Computing Machinery. pp. 15–21. doi:10.1145/3375627.3375814. ISBN   978-1-4503-7110-0.
  2. Hern, Alex (2016-09-28). "'Partnership on AI' formed by Google, DeepMind, Facebook, Amazon, IBM and Microsoft | Technology". The Guardian. Retrieved 2016-09-29.
  3. Waters, Richard (2016-09-29). "AI is 'Next Big Thing' to worry about". Financial Times. Retrieved 2016-09-29.
  4. Bindi, Tas (2016-09-29). "Amazon, Google, Facebook, IBM, and Microsoft form AI non-profit". ZDNet. Retrieved 2016-09-29.
  5. Rubin, Ben Fox; Cheng, Roger (2016-09-29). "The AI Super Friends assemble! (The 3:59, Ep. 115)". CNET. Archived from the original on 2016-09-30. Retrieved 2016-09-29.
  6. "New Partners To Bolster Perspective For Responsible AI". 24 September 2019.
  7. Tilley, Aaron (January 27, 2017). "Why Apple Joined Rivals Amazon, Google, Microsoft In AI Partnership". Forbes. Retrieved 2017-02-02.
  8. "Partnership on AI Announces Executive Director Terah Lyons and Welcomes New Partners - The Partnership on AI". The Partnership on AI. 2017-10-17. Retrieved 2018-09-10.
  9. Taylor, Chloe (2018-10-17). "Baidu becomes the first Chinese firm to join US-led A.I. body". CNBC. Archived from the original on November 3, 2018. Retrieved 2018-10-17.
  10. McGregor, Sean (2020-11-18). "When AI Systems Fail: Introducing the AI Incident Database". partnershiponai.org. Partnership on AI. Retrieved 2020-11-21. Avoiding repeated AI failures requires making past failures known. Therefore, today we introduce a systematized collection of incidents where intelligent systems have caused safety, fairness, or other real-world problems: The AI Incident Database (AIID).
  11. 1 2 Hongo, Hudson (2021-10-01). "PAI Submits Response to NIST's Request for Information on AI Risk Management Framework". Partnership on AI. Retrieved 2023-12-05.
  12. "Rebecca Finlay named as CEO: A Letter from Eric Horvitz, Chair of the Board of Directors". Partnership on AI. 2021-10-26. Retrieved 2022-03-18.
  13. Sosa, Penelope (2023-02-27). "Industry Leaders and Advocates Launch Framework for Responsible Use of AI-Generated Media". Partnership on AI. Retrieved 2023-12-05.
  14. "OSF". osf.io. Retrieved 2023-12-06.
  15. "AI & Media Integrity". Partnership on AI. Retrieved 2023-12-06.
  16. "AI, Labor, and the Economy". Partnership on AI. Retrieved 2023-12-06.
  17. "Fairness, Transparency, and Accountability & ABOUT ML". Partnership on AI. Retrieved 2023-12-06.
  18. "Inclusive Research & Design". Partnership on AI. Retrieved 2023-12-06.
  19. "Safety Critical AI". Partnership on AI. Retrieved 2023-12-06.
  20. Morandín-Ahuerma, Fabio. "OSF". osf.io. doi:10.31219/osf.io/sj2z5 . Retrieved 2023-12-06.
  21. "PAI's Guidance for Safe Foundation Model Deployment". Partnership on AI. Retrieved 2023-12-03.
  22. "Investigating Challenges to Diversity in AI". Partnership on AI. 2021-01-30. Retrieved 2023-12-03.
  23. "Convening Across Industries". Partnership on AI. 2021-01-30. Retrieved 2023-12-03.
  24. "Taking a Methodical Approach to Best Practices". Partnership on AI. 2021-01-30. Retrieved 2023-12-03.
  25. "Inclusive Research & Design". Partnership on AI. Retrieved 2023-12-03.
  26. "AI Incidents Database". Partnership on AI. Retrieved 2023-12-03.
  27. "PAI's 2023 Policy Forum - YouTube". www.youtube.com. Retrieved 2023-12-05.
  28. 1 2 3 4 5 6 7 8 9 10 11 12 13 "Our Team". Partnership on AI. Retrieved 2023-12-05.
  29. "Access Now resigns from the Partnership on AI". Access Now. Retrieved 2023-12-05.