Social machine

Last updated
Early computing machinery was used to establish the US Social Security Administration. As the largest bookkeeping project in history, this would not have been possible without such technology. Early SSA accounting operations.jpg
Early computing machinery was used to establish the US Social Security Administration. As the largest bookkeeping project in history, this would not have been possible without such technology.
Graphical representation of social machines on the Internet that have access to big data in data bases via reference architectures and that communicate with many users in social networks via human language. Social.Machines.Networks.and.Big.Data.png
Graphical representation of social machines on the Internet that have access to big data in data bases via reference architectures and that communicate with many users in social networks via human language.

A social machine is an environment comprising humans and technology interacting and producing outputs or action which would not be possible without both parties present. It can also be regarded as a machine, in which specific tasks are performed by human participants, whose interaction is mediated by an infrastructure (typically, but not necessarily, digital). [1] The growth of social machines has been greatly enabled by technologies such as the Internet, the smartphone, social media and the World Wide Web, by connecting people in new ways. [2]

Contents

Concept

The idea of social machines has been around for a long time, discussed as early as 1846 by Captain William Allen, and also by authors such as Norman Mailer, Gilles Deleuze and Félix Guattari. [2]

Social machines blur the lines between computational processes and input from humans. [3] They often take the form of collaborative online projects which produce web content, such as Wikipedia, [4] citizen science projects like Galaxy Zoo, and even social networking site such as Twitter have also been defined as social machines. [5] However, a social machine does not necessarily produce outcomes which directly affect the individuals or machines involved and an alternative viewpoint states that Social Machines are "rather than being an intentionally engineered piece of software - the substrate of accumulated human cross-system information sharing activities". [6] [7]

Nigel Shadbolt et al. say that the telos of the social machine is specific to its participants, whereas the telos of a platform is independent of its participants’ purposes; the platform is there to facilitate communication. A social machine may also spread across more than one platform, depending on how its participants interact, while a platform like Twitter could host many thousands of social machines. [2]

An academic field investigating the idea has been active [8] since Tim Berners-Lee's book Weaving the web . Social machines are characterised as 'social systems on the Web ... computational entities governed by both computational and social processes'. [9] Tim Berners-Lee and James Hendler expressed some of the underlying scientific challenges with respect to AI research [10] using semantic web technology as a point of departure.

Recent work focuses on the idea that certain social machines can be regarded as autonomous and goal-driven agents, and should be analysed and regulated as such. [11] Nello Cristianini and Teresa Scantamburlo argued that the combination of a human society and an algorithmic regulation forms a social machine. [12] Cristianini's book The Shortcut discusses extensively social machines as a model for many online platforms where participants automatically annotate content during usage, in this way contributing to the overall behaviour of the system. [1]

See also

Related Research Articles

<span class="mw-page-title-main">Semantic Web</span> Extension of the Web to facilitate data exchange

The Semantic Web, sometimes known as Web 3.0, is an extension of the World Wide Web through standards set by the World Wide Web Consortium (W3C). The goal of the Semantic Web is to make Internet data machine-readable.

<span class="mw-page-title-main">Tim Berners-Lee</span> English computer scientist, inventor of the World Wide Web (born 1955)

Sir Timothy John Berners-Lee, also known as TimBL, is an English computer scientist best known as the inventor of the World Wide Web, the HTML markup language, the URL system, and HTTP. He is a professorial research fellow at the University of Oxford and a professor emeritus at the Massachusetts Institute of Technology (MIT).

Text mining, text data mining (TDM) or text analytics is the process of deriving high-quality information from text. It involves "the discovery by computer of new, previously unknown information, by automatically extracting information from different written resources." Written resources may include websites, books, emails, reviews, and articles. High-quality information is typically obtained by devising patterns and trends by means such as statistical pattern learning. According to Hotho et al. (2005) we can distinguish between three different perspectives of text mining: information extraction, data mining, and a knowledge discovery in databases (KDD) process. Text mining usually involves the process of structuring the input text, deriving patterns within the structured data, and finally evaluation and interpretation of the output. 'High quality' in text mining usually refers to some combination of relevance, novelty, and interest. Typical text mining tasks include text categorization, text clustering, concept/entity extraction, production of granular taxonomies, sentiment analysis, document summarization, and entity relation modeling.

<span class="mw-page-title-main">Computational sociology</span> Branch of the discipline of sociology

Computational sociology is a branch of sociology that uses computationally intensive methods to analyze and model social phenomena. Using computer simulations, artificial intelligence, complex statistical methods, and analytic approaches like social network analysis, computational sociology develops and tests theories of complex social processes through bottom-up modeling of social interactions.

Nello Cristianini is a professor of Artificial Intelligence in the Department of Computer Science at the University of Bath.

<span class="mw-page-title-main">James Hendler</span> AI researcher

James Alexander Hendler is an artificial intelligence researcher at Rensselaer Polytechnic Institute, United States, and one of the originators of the Semantic Web. He is a Fellow of the National Academy of Public Administration.

The ESP game is a human-based computation game developed to address the problem of creating difficult metadata. The idea behind the game is to use the computational power of humans to perform a task that computers cannot by packaging the task as a game. It was originally conceived by Luis von Ahn of Carnegie Mellon University. Google bought a license to create its own version of the game in 2006 in order to return better search results for its online images. The license of the data acquired by Ahn's ESP game, or the Google version, is not clear. Google's version was shut down on September 16, 2011, as part of the Google Labs closure in September 2011.

Semantic publishing on the Web, or semantic web publishing, refers to publishing information on the web as documents accompanied by semantic markup. Semantic publication provides a way for computers to understand the structure and even the meaning of the published information, making information search and data integration more efficient.

<span class="mw-page-title-main">Web Science Trust</span> UK Charitable Trust

The Web Science Trust (WST) is a UK Charitable Trust with the aim of supporting the global development of Web science. It was originally started in 2006 as a joint effort between MIT and University of Southampton to formalise the social and technical aspects of the World Wide Web. The trust coordinates a set of international "WSTNet Laboratories" that include academic research groups in the emerging area of Web science.

<span class="mw-page-title-main">Web science</span> Emerging interdisciplinary field

Web science is an emerging interdisciplinary field concerned with the study of large-scale socio-technical systems, particularly the World Wide Web. It considers the relationship between people and technology, the ways that society and technology co-constitute one another and the impact of this co-constitution on broader society. Web Science combines research from disciplines as diverse as sociology, computer science, economics, and mathematics.

Frames are an artificial intelligence data structure used to divide knowledge into substructures by representing "stereotyped situations". They were proposed by Marvin Minsky in his 1974 article "A Framework for Representing Knowledge". Frames are the primary data structure used in artificial intelligence frame languages; they are stored as ontologies of sets.

<span class="mw-page-title-main">Semantic HTML</span> HTML used to reinforce meaning of documents or webpages

Semantic HTML is the use of HTML markup to reinforce the semantics, or meaning, of the information in web pages and web applications rather than merely to define its presentation or look. Semantic HTML is processed by traditional web browsers as well as by many other user agents. CSS is used to suggest its presentation to human users.

<span class="mw-page-title-main">Wendy Hall</span> British computer scientist

Dame Wendy Hall is a British computer scientist. She is Regius Professor of Computer Science at the University of Southampton.

<span class="mw-page-title-main">Nigel Shadbolt</span> Principal of Jesus College, Oxford

Sir Nigel Richard Shadbolt is Principal of Jesus College, Oxford, and Professorial Research Fellow in the Department of Computer Science, University of Oxford. He is chairman of the Open Data Institute which he co-founded with Tim Berners-Lee. He is also a visiting professor in the School of Electronics and Computer Science at the University of Southampton. Shadbolt is an interdisciplinary researcher, policy expert and commentator. His research focuses on understanding how intelligent behaviour is embodied and emerges in humans, machines and, most recently, on the Web, and has made contributions to the fields of Psychology, Cognitive science, Computational neuroscience, Artificial Intelligence (AI), Computer science and the emerging field of Web science.

<span class="mw-page-title-main">David De Roure</span> English computer scientist

David Charles De Roure is an English computer scientist who is a professor of e-Research at the University of Oxford, where he is responsible for Digital Humanities in The Oxford Research Centre in the Humanities (TORCH), and is a Turing Fellow at The Alan Turing Institute. He is a supernumerary Fellow of Wolfson College, Oxford, and Oxford Martin School Senior Alumni Fellow.

Computational social science is an interdisciplinary academic sub-field concerned with computational approaches to the social sciences. This means that computers are used to model, simulate, and analyze social phenomena. It has been applied in areas such as computational economics, computational sociology, computational media analysis, cliodynamics, culturomics, nonprofit studies. It focuses on investigating social and behavioral relationships and interactions using data science approaches, network analysis, social simulation and studies using interactive systems.

<span class="mw-page-title-main">Jeni Tennison</span>

Jenifer Fays Alys Tennison is a British software engineer and consultant who co-chairs the data governance working group within the Global Partnership on Artificial Intelligence (GPAI). She also serves on the board of directors of Creative Commons, the Global Partnership for Sustainable Development Data (GPSDD) and the information law and policy centre of the School of Advanced Study (SAS) at the University of London. She was previously Chief Executive Officer (CEO) of the Open Data Institute (ODI).

In natural language processing (NLP), a word embedding is a representation of a word. The embedding is used in text analysis. Typically, the representation is a real-valued vector that encodes the meaning of the word in such a way that the words that are closer in the vector space are expected to be similar in meaning. Word embeddings can be obtained using language modeling and feature learning techniques, where words or phrases from the vocabulary are mapped to vectors of real numbers.

Health Web Science (HWS) is a sub-discipline of Web Science that examines the interplay between health sciences, health and well-being, and the World Wide Web. It assumes that each domain influences the others. HWS thus complements and overlaps with Medicine 2.0. Research has uncovered emergent properties that arise as individuals interact with each other, with healthcare providers and with the Web itself.

<span class="mw-page-title-main">Daniel Weitzner</span> Scientist

Daniel J. Weitzner is the director of the MIT Internet Policy Research Initiative and principal research scientist at the Computer Science and Artificial Intelligence Lab CSAIL. He teaches Internet public policy in MIT's Computer Science Department. His research includes development of accountable systems architectures to enable the Web to be more responsive to policy requirements.

References

  1. 1 2 Cristianini, Nello (2023). The shortcut : why intelligent machines do not think like us (First ed.). Boca Raton. ISBN   978-1-003-33581-8. OCLC   1352480147.{{cite book}}: CS1 maint: location missing publisher (link)
  2. 1 2 3 Shadbolt, Nigel; O'Hara, Kieron; De Roure, David; Hall, Wendy (2019), The Theory and Practice of Social Machines, Springer, ISBN   978-3-030-10888-5
  3. Shadbolt, Nigel R.; Daniel A. Smith; Elena Simperl; Max Van Kleek; Yang Yang; Wendy Hall (2013). "Towards a Classification Framework for Social Machines" (PDF). WWW 2013 Companion. Retrieved 25 May 2014.
  4. Berners-Lee, Tim; J. Hendler (2009). "From the Semantic Web to social machines: A research challenge for AI on the World WideWeb" (PDF). Artificial Intelligence. 174 (2): 156–161. doi: 10.1016/j.artint.2009.11.010 . Retrieved 25 May 2014.
  5. Meira, Silvio R. L.; Buregio, Vanilson A. A.; Nascimento, Leandro M.; Figueiredo, Elaine; Neto, Misael; Encarnacao, Bruno; Garcia, Vinicius Cardoso (2011). "The Emerging Web of Social Machines". 2011 IEEE 35th Annual Computer Software and Applications Conference. pp. 26–27. arXiv: 1010.3045 . doi:10.1109/COMPSAC.2011.12. ISBN   978-1-4577-0544-1. S2CID   954471.
  6. Luczak-Roesch, M.; Tinati, R.; O'Hara, K.; Shadbolt, N. (2015). "Socio-technical Computation". Proceedings of the 18th ACM Conference Companion on Computer Supported Cooperative Work & Social Computing (PDF). ACM. pp. 139–142. doi:10.1145/2685553.2698991. ISBN   9781450329460. S2CID   22103905.{{cite book}}: CS1 maint: date and year (link)
  7. Luczak-Roesch, M.; Tinati, R.; Shadbolt, N. (2015). "When Resources Collide: Towards a Theory of Coincidence in Information Spaces". Proceedings of the 24th International Conference on World Wide Web (PDF). ACM. pp. 1137–1142. doi:10.1145/2740908.2743973. ISBN   9781450334730. S2CID   17495801.{{cite book}}: |journal= ignored (help)
  8. Shadbolt, N.; Kleek, M. Van; Binns, R. (2016-04-01). "The rise of social machines: The development of a human/digital ecosystem". IEEE Consumer Electronics Magazine. 5 (2): 106–111. doi:10.1109/MCE.2016.2516179. ISSN   2162-2248. S2CID   24896885.
  9. "About". SOCIAM. Retrieved 25 May 2014.
  10. Hendler, J.; Berners-Lee, T. (2010-02-10). ""From the Semantic Web to Social Machines"". AI Journal. 2 (174). doi: 10.1016/j.artint.2009.11.010 .
  11. Cristianini, Nello; Scantamburlo, Teresa; Ladyman, James (4 October 2021). "The social turn of artificial intelligence" (PDF). AI & Society. 38: 89–96. doi:10.1007/s00146-021-01289-8. S2CID   244180663.
  12. Cristianini, Nello; Scantamburlo, Teresa (8 October 2019). "On social machines for algorithmic regulation". AI & Society. 35 (3): 645–662. arXiv: 1904.13316 . Bibcode:2019arXiv190413316C. doi:10.1007/s00146-019-00917-8. ISSN   1435-5655. S2CID   140233845.

Further reading