This article has multiple issues. Please help improve it or discuss these issues on the talk page . (Learn how and when to remove these template messages)
|
Government crowdsourcing is a form of crowdsourcing employed by governments to better leverage their constituents' collective knowledge and experience. [1] It has tended to take the form of public feedback, project development, or petitions in the past, but has grown to include public drafting of bills and constitutions, among other things. [2] This form of public involvement in the governing process differs from older systems of popular action, from town halls to referendums, in that it is primarily conducted online or through a similar IT medium. [2]
Various thinkers, including but not limited to Daren Brabham, Beth Noveck, and Helene Landemore, have each presented their own definitions of what crowdsourcing as a whole, and government crowdsourcing by extension, necessarily entails, but there has been no consensus thus far. [3] [4] Governments which have adopted crowdsourcing as a method of information gathering, policy guidance, and in some cases a vital part of the lawmaking process include Brazil, Finland, Iceland, Egypt, Tunisia, and the United States, among many others. [3] [5] [6] [7] [8]
Though, in the past, direct democracy employed many of the same mechanisms as government crowdsourcing, and indeed could resemble it at times, the objectives direct democracy tended to pursue and the methods it used set it apart. Crowdsourcing requires a specific goal, rewards to be gained by both the crowd and the government, and an open request for anyone to participate, as well as to be conducted through the internet or some other IT medium. [9] Programs which truly resembled pre-Internet government crowdsourcing did not emerge until the 17th and 18th centuries. These were government-sponsored competitions such as the Alkali prize, which lead to the development of the Leblanc Process, and the Longitude Prize.
Other early precursors to government crowdsourcing either followed the same model of reaching out in search of an expert or inventor capable of solving the problem at hand, via what Brabham refers to as the "broadcast-search model". The American government distributed large tasks, such as mapping wind and current patterns along trade routes among American sailors, delegating the discovery of existing knowledge to a crowd in a model similar to what is now referred to as "distributed human intelligence tasking". [10] [11]
Finally, the French Cahiers de doléances were the most directly governmental example of pre-internet government crowdsourcing. In the lead up to the French Revolution, the three estates listed their grievances and made suggestions to improve the government. [12] These problems, complaints, and proposals were considered and debated through the sessions of the Estates general. [13] Similar efforts to hear the grievances and suggestions of subjects, and later constituents, can be seen at the heart of various political institutions around the world, such as town halls in the United States, petitions in the United Kingdom and beyond.
Today, government crowdsourcing follows similar models, albeit utilizing the internet, mobile phones and other IT mediums.[ citation needed ] Around the world, governments continue to use competitions, delegate large tasks, and reach out to their constituents for feedback both on specific bills and for guidance on various policies. They have also begun, in countries as different as Brazil, British Columbia, and Finland, to not only draw feedback, expertise, and ideas from the crowd, but also specific provisions and wordings for legislation and even constitutions. [14] [7] These initiatives, which give participating constituents greater leeway to both develop their ideas, deliberate, and be heard, mirroring what Brabham describes as "peer-vetted creative crowdsourcing". [15]
The established models persist, but attempts at crowd law and crowdsourced constitutions have met with mixed results. In some cases, the crowdsourcing platform has been purely suggestive, as in British Columbia, and their proposal ultimately put to a vote. Generally, once those crowdsourced suggestions entered the political system, they were either voted down or buried under mountains of procedural delay. [5] [16] In other rarer cases, the efforts succeeded and the crowdsourced laws were adopted, though in somewhat amended form, as happened with the Brazilian Internet Bill of Rights and in the Constitutional Convention (Ireland). Other efforts to crowdsource government are still in process, or have only been tested on a limited scale, leaving their long-term results still uncertain. [17]
In several post-revolutionary moments, notably those of Tunisia, Iceland, and Egypt, the new governments have also used crowdsourcing as a means of securing legitimacy and addressing the issues which brought them into power in the first place. This method of constitution making has been met with a limited level of success. The Icelandic constitution, crowdsourced in the aftermath of the 2009 Icelandic financial crisis protests, utilized a multi-layered crowdsourcing structure, with an elected council at the top synthesizing suggestions into the new constitution, a popular forum for the crowd to make their voiced heard, and a Constitutional committee to organize the other parts. [5] By contrast, the Egyptian and Tunisian constitutional processes involved the crowd much less, with an official constitutional assembly composed of traditional political elites who simply received feedback on their draft clauses online.
Today, there is still no definitive definition of crowdsourcing, and even less agreement as to what aspects of participatory democracy fit under that description. Most academics and writers tend to create their own definitions. The most expansive definition of crowdsourcing, drawn from a survey of thousands of papers on the topic, describes it as something for which: (a) there is a clearly defined crowd; (b) there exists a task with a clear goal; (c) the recompense received by the crowd is clear; (d) the crowd-sourcer is clearly identified; (e) the compensation to be received by the crowd-sourcer is clearly defined; (f) it is an online assigned process of participative type; (g) it uses an open call of variable extent; (h) it uses the internet [9]
Other definitions are less exhaustive and more case-specific. Helene Landemore, following a similar line of thought as Brabham, defines crowdsourcing as an online problem-solving and production model in which an undefined crowd helps to complete a task by submitting knowledge, information, or talent. In its unrefined form, there is no accountability mechanism, and the crowd, unlike contractors in an outsourcing system, is not vetted. [5] She further distinguishes it from Wikipedia-esque "commons-based peer production", noting that crowdsourcing generally takes the form of individuals commenting and making suggestions in a vacuum as opposed to deliberating, discussing, and collaborating with one another. [18] [5] [19]
Brahbam also makes a subtle distinction, which makes his definition substantially more restrictive. For an aspect of participatory government to be crowdsourcing, it cannot be purely a government-driven sequence of responses from the crowd. Both sides must engage in direction and deliberation on the project. [20] Under this definition, certain elements of governmental participation, such as the Peer to Patent project are not crowdsourcing at all, but rather similar processes which have been agglomerated under a new and trendy name.
Mabhoudi, on the other hand, defines it purely in terms of constitution making. His conception of crowdsourcing consists of posting a draft constitution online, utilizing both official websites and social media pages. Those platforms were used for feedback, commentary, recommendations, and possibly final approval. Left out of his definition, but noted in his description of the process of Egyptian constitution making is the use of committees to document, digest, and synthesize the volumes of online feedback into a more readable form, which served an indispensable role in the success, however temporary, of the crowdsourcing process. [21]
Noveck refers to government crowdsourcing as "collaborative democracy". [22] She defines that as a process of using technology to improve government outcomes by soliciting experience from groups of self-selected peers working together in groups of open networks. [22] Her version of government crowdsourcing more closely resembles the government competitions of the 17th and 18th centuries, or Amazon's Mechanical Turk, than the more participatory model favored by other thinkers. [23] That is, crowds gathered to work on an agenda set out by the government, [24] pooling their diverse expertise to cover every possible field of information.
Open Opportunities is a government wide program offering professional development opportunities to current federal employees and internships to students. The program facilitates collaboration and knowledge sharing across the Federal Government. [25]
Open Opportunities is for federal employees looking to gain additional experience, and students looking for internships. The program offers a wide variety of real world projects that promote experiential learning. [25]
The Virtual Student Foreign Service is a crowd-work and eIntern program for college students. [26]
The Office of eDiplomacy is developing an internal crowdsourcing platform called CrowdWork that will facilitate collaborative work worldwide. Any office or mission will be able to post tasks online and any State Department employee with the requisite skills will be able to respond and complete the task. This creates an internal marketplace for foreign affairs work and matches State Department opportunities and requirements with untapped skills and experience. The platform is currently under development with an anticipated launch in December 2013. This crowd-work platform will be developed as part of an innovation toolkit for the U.S. Government. [27] This project is developed through a partnership between the State Department office of eDiplomacy, [28] the White House Office of Science and Technology Policy, the General Services Administration and the State Department Office of the Director General. The crowd work element of the innovation toolkit will be developed as an open-source platform for public use. [29]
The State Department Sounding Board is an internal ideation tool for employees to suggest improvements to the Department.
The Humanitarian Information Unit (HIU), [30] a division within the Office of the Geographer and Global Issues at the U.S. Department of State, is working to increase the availability of spatial data in areas experiencing humanitarian emergencies. Built from a crowdsourcing model, the new "Imagery to the Crowd" process publishes high-resolution commercial satellite imagery, purchased by the United States Government, in a web-based format that can be easily mapped by volunteers. [31] The digital map data generated by the volunteers are stored in a database maintained by OpenStreetMap (OSM), a UK-registered non-profit foundation, under a license that ensures the data are freely available and open for a range of uses.
Inspired by the success of the OSM mapping effort after the 2010 Haiti earthquake, the Imagery to the Crowd process harnesses the combined power of satellite imagery and the volunteer mapping community to help aid agencies provide informed and effective humanitarian assistance, and plan recovery and development activities. The HIU partners with the Humanitarian OpenStreetMap Team (HOT) on many of the Imagery to the Crowd projects. HOT provides volunteer support and access to its micro-tasking platform, the OSM Tasking Manager, which coordinates volunteer efforts by breaking down large mapping tasks into smaller areas that can be digitized in 45–60 minutes. A 5-minute Imagery to the Crowd Ignite talk is available. [32]
The Bureau of Education and Cultural Affairs partnered with RocketHub on the Alumni Engagement and Innovation Fund (AEIF) 2.0 which helps exchange program alumni crowdsource funding for their innovative projects.
The Special Advisor for Civil Society and Emerging Democracies created Diplomacy Lab as a way for the Department to crowdsource longer-term projects from the academic community.
Human Resource's Flex Connect program facilitates crowdsourcing of talent from across the Department.
eDiplomacy is also piloting a State Department GitHub Account to collaborate on open source software projects.
The Bureau of Arms Control, Verification and Compliance also ran a crowdsourced Arms Control Challenge examining how technology could aid arms control efforts.
In June 2012, USAID launched the Agency's first-ever crowdsourcing initiative to pinpoint the location of USAID Development Credit Authority (DCA) loan data and make the dataset publicly available. Crowdsourcing is a distributed problem-solving process whereby tasks are outsourced to a network of people known as "the crowd".
The engagement of the Crowd was an innovative way to process data and increase the transparency of the Agency. Visualizing where USAID enhances the capacity of the private sector can signal new areas for potential collaboration with host countries, researchers, development organizations, and the public. A case study explains the organizational, legal, and technical steps behind making these data open. [33]
CDCology [34] is a way for CDC staff to post unclassified, one-minute to one-day long (micro)tasks that can be solved by undergraduate, graduate, and post-graduate student volunteers. This expands the agency's workforce and relieves staff to focus on in-depth assignments. In return, universities can offer students short challenges directly impacting national initiatives and gain experience with government work. This allows students to bring fresh ideas to smaller projects and they can add "micro-volunteering for CDC" to their résumé. [35]
Pilot project launched by the Administrator's office called Skills Marketplace allowing micro-details [ further explanation needed ] to other offices and projects. [36]
Challenge.gov is an online challenge platform administered by the U.S. General Services Administration (GSA) in partnership with Challenge Post that empowers the U.S. Government and the public to bring the best ideas and top talent to bear on our nation's most pressing challenges. This platform is the latest milestone in the Administration's commitment to use prizes and challenges to promote innovation. [37]
Idea Factory empowers the Transportation Security Administration's large and dispersed workforce to submit and collaborate on innovative ideas to improve TSA and keep the nation's transportation systems secure. [38]
Digital Volunteers is a crowd-work platform for transcribing historical documents. [39]
eMammal is a crowdsourcing collection of images and data on N American mammal populations [40] [41]
The National Archives uses digital volunteers to identify signatures in documents and index the archive. [42]
Remember Me is a project to post 1100 pictures of children displaced during the Jewish Holocaust to identify these children, piece together information about their wartime and postwar experiences, and facilitate renewed connections among these young survivors, their families, and other individuals who were involved in their care during and after the war. [43]
The World Memory Project aims to digitize the records of victims of the Jewish Holocaust. [44]
CoCreate is a crowdsourced effort to identify and solve real life soldier challenges.
The US Navy crowdsourced solutions to Somali Piracy via its Massive Multiplayer Online Wargame Leveraging the Internet. [45] It is a message-based game to encourage innovative thinking by many people, connected via the Web. It has been used to study a number of topics, such as how the Navy can prepare for the future of energy......starting in 2021 and beyond
The United States Patent and Trademark Office (USPTO) has been working to integrate crowdsourcing into the patent process in a number of ways. The first key step was the introduction of the Peer to Patent program at the application review stage. Peer to Patent was designed to allow members of the public with relevant technical skills and information to submit information useful to the patent examiner when assessing the claims of pending patent applications. The first pilot Peer to Patent program, limited to software and business methods applications, was launched in 2007 and remained open through 2009. [46] After reviewing the results, a second expanded pilot was launched in 2010, extending coverage to include biotechnology, bioinformatics, telecommunications, and speech recognition. [47]
The USPTO was further encouraged to implement crowdsourcing by the Obama Administration's Executive Action on Crowdsourcing Prior Art. [48] The order was created with the goal of improving patent quality overall and fostering innovation by integrating new ways for companies, experts, and the general public to find and submit "prior art" or evidence needed to ascertain if an invention is novel. [49]
Having tested crowdsourcing during the application review process, the USPTO turned to explore applications for crowdsourcing during post-grant patent review. To do this, the USPTO consulted experts in the government and private sector to share their expertise through two roundtable discussions. The Roundtable on the Use of Crowdsourcing and Third-Party Preissuance Submissions to Identify Relevant Prior Art, took place April 10, 2014. Presenters at the roundtable included Andrea Casillas and Christopher Wong sharing their experience with Peer to Patent; Micah Siegel from Ask Patents, a patent crowdsourcing project through stack exchange; Pedram Sameni from Patexia, a crowdsourcing platform focused on patent research and prior art searching; and Cheryl Milone of Article One Partners. [50]
The USPTO held their second roundtable on the use of crowdsourcing to identify relevant prior art on December 2, 2014. This roundtable focused on two key questions: (1) how the USPTO can utilize crowdsourcing tools to obtain relevant prior art in order to enhance the quality of examination and issued patents; and (2) ways the USPTO can leverage existing private sector solutions for the electronic receipt and hosting of crowdsourced materials as a means to provide prior art to examiners. [51] Speakers at the roundtable included Matt Levy (Computer & Communications Industry Association), Mark Nowotarski (Markets, Patents & Alliances LLC), Cheryl Milone (Article One Partners), and Pedram Sameni (Patexia Inc.).
The USPTO continues to consider additional applications for crowdsourcing following the roundtables and public comments submitted in response. As part of that goal, they are working with Christopher Wong, who was appointed to be the Presidential Innovation Fellow supporting crowdsourcing and patent reform initiatives for the USPTO. [52]
The Patient Feedback Challenge is about getting the NHS to use feedback from patients to improve services, by spreading the best approaches already out there. Ideas are crowdsourced, and then bid on by NHS organizations. [53]
Since 2010 the FCO has made the Human Rights and Democracy Report available online and available for markup by NGOs, policy makers, academics and the general public. The comments are forwarded to the relevant policy teams for evaluation and to respond to accordingly. [54]
Topcoder is a crowdsourcing company with an open global community of designers, developers, data scientists, and competitive programmers. Topcoder pays community members for their work on the projects and sells community services to corporate, mid-size, and small-business clients. Topcoder also organizes the annual Topcoder Open tournament and a series of smaller regional events.
Electronic participation (e-participation) refers to the use of ICT in facilitating citizen participation in government-related processes, encompassing areas such as administration, service delivery, decision-making, and policy-making. As such, e-participation shares close ties with e-government and e-governance participation. The term's emergence aligns with the digitization of citizen interests and interactions with political service providers, primarily due to the proliferation of e-government.
Crowdsourcing involves a large group of dispersed participants contributing or producing goods or services—including ideas, votes, micro-tasks, and finances—for payment or as volunteers. Contemporary crowdsourcing often involves digital platforms to attract and divide work between participants to achieve a cumulative result. Crowdsourcing is not limited to online activity, however, and there are various historical examples of crowdsourcing. The word crowdsourcing is a portmanteau of "crowd" and "outsourcing". In contrast to outsourcing, crowdsourcing usually involves less specific and more public groups of participants.
Crowdreviewing is the practice of gathering opinion or feedback from a large number of people, typically via the internet or an online community; a portmanteau of "crowd" and "reviews". Crowdreviewing is also often viewed as a form of crowd voting which occurs when a website gathers a large group's opinions and judgment. The concept is based on the principles of crowdsourcing and lets users submit online reviews to participate in building online metrics that measure performance. By harnessing social collaboration in the form of feedback individuals are generally able to form a more informed opinion.
The involvement of the public in patent examination is used in some forms to help identifying relevant prior art and, more generally, to help assessing whether patent applications and inventions meet the requirements of patent law, such as novelty, inventive step or non-obviousness, and sufficiency of disclosure.
The Peer To Patent project is an initiative that seeks to assist patent offices in improving patent quality by gathering public input in a structured, productive manner. Peer To Patent is the first social-software project directly linked to decision-making by the federal government.
Beth Simone Noveck is a Professor at Northeastern University and the 1st Chief AI Strategist for the State of New Jersey. She previously served as founding Chief Innovation Officer of New Jersey. At Northeastern, she directs the Burnes Center for Social Change and its partner project, The Governance Lab. She is also affiliated faculty with the Institute for Experiential AI. She is the author of Solving Public Problems: How to Fix our Government and Change Our World, Smart Citizens, Smarter State: The Technologies of Expertise and the Future of Government, Wiki Government: How Technology Can Make Government Better, Democracy Stronger, and Citizens More Powerful, and co-editor of the State of Play: Law and Virtual Worlds.
Figure Eight was a human-in-the-loop machine learning and artificial intelligence company based in San Francisco.
Citizen sourcing is the government adoption of crowdsourcing techniques for the purposes of (1) enlisting citizens in the design and execution of government services and (2) tapping into the citizenry's collective intelligence for solutions and situational awareness. Applications of citizen sourcing include:
Article One Partners (AOP) is an online prior art search and intellectual property research crowdsourcing community. AOP was acquired by RWS Group in October 2017 and the AOP Connect crowdsourcing platform is now part of the IP Research group within RWS. RWS IP Research provides crowdsourced prior-art-search services by utilizing an online research community. The company's President stated that, as of June 30, 2018, AOP comprises "more than 40,000 in over 170 countries."
Jade Magnet was an online crowdsourcing platform for creative and marketing support services based in Bangalore, India. It was founded in 2009 by Sitashwa Srivastava and Manik Kinra. The company has white label partnerships in Qatar as Mixilion and in Singapore as id8on.
Crisis mapping is the real-time gathering, display and analysis of data during a crisis, usually a natural disaster or social/political conflict. Crisis mapping projects usually allows large numbers of people, including the public and crisis responders, to contribute information either remotely or from the site of the crisis. One benefit of the crisis mapping method over others is that it can increase situational awareness, since the public can report information and improve data management.
Crowdsourcing software development or software crowdsourcing is an emerging area of software engineering. It is an open call for participation in any task of software development, including documentation, design, coding and testing. These tasks are normally conducted by either members of a software enterprise or people contracted by the enterprise. But in software crowdsourcing, all the tasks can be assigned to or are addressed by members of the general public. Individuals and teams may also participate in crowdsourcing contests.
Civic technology, or civic tech, enhances the relationship between the people and government with software for communications, decision-making, service delivery, and political process. It includes information and communications technology supporting government with software built by community-led teams of volunteers, nonprofits, consultants, and private companies as well as embedded tech teams working within government.
Mapillary is a service for sharing crowdsourced geotagged photos, developed by remote company Mapillary AB, based in Malmö, Sweden. Mapillary was launched in 2013 and acquired by Meta Platforms in 2020. It offers street level imagery similar to Google Street View.
Crowdsourced Testing is a crowdsourcing platform which provides functional, localization, usability and Beta testing through crowdsourcing.
Patexia Inc. is a privately held intellectual property (IP) company based in Santa Monica, California, U.S. The company was founded in 2010 with the mission to enhance transparency and efficiency in the IP field through a leveraging of the knowledge of an IP-based online community of researchers, attorneys, and stakeholders—described by the company as a “multidisciplinary social network”—for the purpose of information crowdsourcing. In addition, the company combines patent and litigation databases to provide analytical tools regarding the IP field, including the details of attorneys, law firms, companies, and examiners, for its community members.
Crowdmapping is a subtype of crowdsourcing by which aggregation of crowd-generated inputs such as captured communications and social media feeds are combined with geographic data to create a digital map that is as up-to-date as possible on events such as wars, humanitarian crises, crime, elections, or natural disasters. Such maps are typically created collaboratively by people coming together over the Internet.
Wiki surveys or wikisurveys are a software-based survey method with similarity to how wikis evolve through crowdsourcing. In essence, they are surveys that allow participants to create the questions that are being asked. As participants engage in the survey they can either vote on a survey question or create a survey question. A single open-ended prompt written by the creator of the survey determines the topic the questions should be on. The first known implementation of a wiki survey was in 2010, and they have been used since then for a variety of purposes such as facilitating deliberative democracy, crowdsourcing opinions from experts and figuring out common beliefs on a given topic. A notable usage of wiki surveys is in Taiwan's government system, where citizens can participate in crowdsourced lawmaking through Polis wiki surveys.