Value sensitive design

Last updated

Value sensitive design (VSD) is a theoretically grounded approach to the design of technology that accounts for human values in a principled and comprehensive manner. [1] [2] VSD originated within the field of information systems design [3] and human-computer interaction [4] to address design issues within the fields by emphasizing the ethical values of direct and indirect stakeholders. It was developed by Batya Friedman and Peter Kahn at the University of Washington starting in the late 1980s and early 1990s. Later, in 2019, Batya Friedman and David Hendry wrote a book on this topic called "Value Sensitive Design: Shaping Technology with Moral Imagination". [5] Value Sensitive Design takes human values into account in a well-defined matter throughout the whole process. [6] Designs are developed using an investigation consisting of three phases: conceptual, empirical and technological. [7] These investigations are intended to be iterative, allowing the designer to modify the design continuously. [8]

Contents

The VSD approach is often described as an approach that is fundamentally predicated on its ability to be modified depending on the technology, value(s), or context of use. [9] [10] Some examples of modified VSD approaches are Privacy by Design which is concerned with respecting the privacy of personally identifiable information in systems and processes. [11] Care-Centered Value Sensitive Design (CCVSD) proposed by Aimee van Wynsberghe is another example of how the VSD approach is modified to account for the values central to care for the design and development of care robots. [12]

Design process

VSD uses an iterative design process that involves three types of investigations: conceptual, empirical and technical. Conceptual investigations aim at understanding and articulating the various stakeholders of the technology, as well as their values and any values conflicts that might arise for these stakeholders through the use of the technology. Empirical investigations are qualitative or quantitative design research studies used to inform the designers' understanding of the users' values, needs, and practices. Technical investigations can involve either analysis of how people use related technologies, or the design of systems to support values identified in the conceptual and empirical investigations. [13] Friedman and Hendry account seventeen methods, including their main purpose, an overview of its function as well as key references: [5]

  1. Stakeholder Analysis (Purpose: Stakeholder identification and legitimation): Identification of individuals, groups, organizations, institutions, and societies that might reasonably be affected by the technology under investigation and in what ways. Two overarching stakeholder categories: (1) those who interact directly with the technology, direct stakeholders; and (2) those indirectly affected by the technology, indirect stakeholders. [14] [15] [16] [17]
  2. Stakeholder Tokens (Purpose: Stakeholder identification and interaction): Playful and versatile toolkit for identifying stakeholders and their interactions. Stakeholder tokens facilitate identifying stakeholders, distinguishing core from peripheral stakeholders, surfacing excluded stakeholders, and articulating relationships among stakeholders. [18]
  3. Value Source Analysis (Purpose: Identify value sources): Distinguish among the explicitly supported project values, designers’ personal values, and values held by other direct and indirect stakeholders. [19]
  4. Co-evolution of Technology and Social Structure (Purpose: Expand design space): Expanding the design space to include social structures integrated with technology may yield new solutions not possible when considering the technology alone. As appropriate, engage with the design of both technology and social structure as part of the solution space. Social structures may include policy, law, regulations, organizational practices, social norms, and others. [20] [21]
  5. Value Scenario (Purpose: Values representation and elicitation): Narratives, comprising stories of use, intended to surface human and technical aspects of technology and context. Value scenarios emphasize implications for direct and indirect stakeholders, related key values, widespread use, indirect impacts, longer-term use, and similar systemic effects. [15] [16] [21]
  6. Value Sketch (Purpose: Values representation and elicitation): Sketching activities as a way to tap into stakeholders’ non-verbal understandings, views, and values about a technology. [22] [23]
  7. Value-oriented Semi-structured Interview (Purpose: Values elicitation): Semi-structured interview questions as a way to tap into stakeholders’ understandings, views and values about a technology. Questions typically emphasize stakeholders’ evaluative judgments (e.g., all right or not all right) about a technology as well as rationale (e.g., why?). Additional considerations introduced by the stakeholder are pursued. [24] [10] [19] [25] [16] [26]
  8. Scalable Information Dimensions (Purpose: Values elicitation): Sets of questions constructed to tease apart the impact of pervasiveness, proximity, granularity of information, and other scalable dimensions. Can be used in interview or survey formats. [24] [14] [27]
  9. Value-oriented Coding Manual (Purpose: Values analysis): Hierarchically structured categories for coding qualitative responses to the value representation and elicitation methods. Coding categories are generated from the data and a conceptualization of the domain. Each category contains a label, definition, and typically up to three sample responses from empirical data. Can be applied to oral, written, and visual responses.[ dead link ]
  10. Value-oriented Mockup, Prototype or Field Deployment (Purpose: Values representation and elicitation): Development, analysis, and co-design of mockups, prototypes and field deployments to scaffold the investigation of value implications of technologies that are yet to be built or widely adopted. Mock-ups, prototypes or field deployments emphasize implications for direct and indirect stakeholders, value tensions, and technology situated in human contexts. [25] [28] [29] [16] [30]
  11. Ethnographically Informed Inquiry regarding Values and Technology (Purpose: Values, technology and social structure framework and analysis): Framework and approach for data collection and analysis to uncover the complex relationships among values, technology and social structure as those relationships unfold. Typically involves indepth engagement in situated contexts over longer periods of time. [31]
  12. Model for Informed Consent Online (Purpose: Design principles and values analysis): Model with corresponding design principles for considering informed consent in online contexts. The construct of informed encompasses disclosure and comprehension; that of consent encompasses voluntariness, competence, and agreement. Furthermore, implementations of informed consent. [32]
  13. Value Dams and Flows (Purpose: Values analysis): Analytic method to reduce the solution space and resolve value tensions among design choices. First, design options that even a small percentage of stakeholders strongly object to are removed from the design space—the value dams. Then of the remaining design options, those that a good percentage of stakeholders find appealing are foregrounded in the design—the value flows. Can be applied to the design of both technology and social structures. [16] [21] [29]
  14. Value Sensitive Action-Reflection Model (Purpose: Values representation and elicitation): Reflective process for introducing value sensitive prompts into a co-design activity. Prompts can be designer or stakeholder generated. [30]
  15. Multi-lifespan timeline (Purpose: Priming longer-term and multi-generational design thinking): Priming activity for longer-term design thinking. Multi-lifespan timelines prompt individuals to situate themselves in a longer time frame relative to the present, with attention to both societal and technological change. [33]
  16. Multi-lifespan co-design (Purpose: Longer-term design thinking and envisioning): Co-design activities and processes that emphasize longer-term anticipatory futures with implications for multiple and future generations. [33]
  17. Envisioning Cards (Purpose: Value sensitive design toolkit for industry, research, and educational practice): Value sensitive envisioning toolkit. A set of 32 cards, the Envisioning Cards build on four criteria: stakeholders, time, values, and pervasiveness. Each card contains on one side a title and an evocative image related to the card theme; on the flip side, the envisioning criterion, card theme, and a focused design activity. Envisioning Cards can be used for ideation, co-design, heuristic critique, and evaluation. [34] [35] [30]

Criticisms

VSD is not without its criticisms. Two commonly cited criticisms are critiques of the heuristics of values on which VSD is built. [36] [37] These critiques have been forwarded by Le Dantec et al. [38] and Manders-Huits. [39] Le Dantec et al. argue that formulating a pre-determined list of implicated values runs the risk of ignoring important values that can be elicited from any given empirical case by mapping those value a priori. [38] Manders-Huits instead takes on the concept of ‘values’ itself with VSD as the central issue. She argues that the traditional VSD definition of values as “what a person or group of people consider important in life” is nebulous and runs the risk of conflating stakeholders preferences with moral values. [39]

Wessel Reijers and Bert Gordijn have built upon the criticisms of Le Dantec et alia and Manders-Huits that the value heuristics of VSD are insufficient given their lack of moral commitment. [37] They propose that a heuristic of virtues stemming from a virtue ethics approach to technology design, mostly influenced by the works of Shannon Vallor, provides a more holistic approach to technology design. Steven Umbrello has criticized this approach arguing that not only can the heuristic of values be reinforced [40] but that VSD does make moral commitments to at least three universal values: human well-being, justice and dignity. [36] [5] Batya Friedman and David Hendry, in "Value Sensitive Design: Shaping Technology with Moral Imagination", argue that although earlier iterations of the VSD approach did not make explicit moral commitments, it has since evolved over the past two decades to commit to at least those three fundamental values. [5]

VSD as a standalone approach has also been criticized as being insufficient for the ethical design of artificial intelligence. [41] This criticism is predicated on the self-learning and opaque artificial intelligence techniques like those stemming from machine learning and, as a consequence, the unforeseen or unforeseeable values or disvalues that may emerge after the deployment of an AI system. Steven Umbrello and Ibo van de Poel propose a modified VSD approach that uses the Artificial Intelligence for Social Good (AI4SG) [42] factors as norms to translate abstract philosophical values into tangible design requirements. [43] What they propose is that full-lifecycle monitoring is necessary to encourage redesign in the event that unwanted values manifest themselves during the deployment of a system.

See also

Related Research Articles

The Stanford Behavior Design Lab is a research organization advancing behavior change methods and models based at Stanford University. Founded in 1998 and directed by B. J. Fogg, the Behavior Design Lab is a team of Stanford students, recent graduates, and quantitative researchers who study factors that impact human behavior, and conduct IRB research. The team is the global authority in a new and systematic way to design for behavior change, an approach called “Behavior Design”. The Lab manager is Tanna Drapkin.

<span class="mw-page-title-main">WIMP (computing)</span> Style of human-computer interaction

In human–computer interaction, WIMP stands for "windows, icons, menus, pointer", denoting a style of interaction using these elements of the user interface. Other expansions are sometimes used, such as substituting "mouse" and "mice" for menus, or "pull-down menu" and "pointing" for pointer.

Robotic pets are artificially intelligent machines that are made to resemble actual pets. While the first robotic pets produced in the late 1990s, were not too advanced, they have since grown technologically. Many now use machine learning, making them much more realistic. Most consumers buy robotic pets with the aim of getting similar companionship that real pets offer, without some of the drawbacks that come with caring for live animals. The pets on the market currently have a wide price range, from the low hundreds into the several thousands of dollars. Multiple studies have been done to show that we treat robotic pets in a similar way as actual pets, despite their obvious differences. However, there is some controversy regarding how ethical using robotic pets is, and whether or not they should be widely adopted in elderly care.

The Special Interest Group on Computer–Human Interaction (SIGCHI) is one of the Association for Computing Machinery's special interest groups which is focused on human–computer interactions (HCI).

ACM Interactions magazine is a publication covering a number of related worlds, offering content to educate and inspire designers, providing viewpoints related to culture and anthropology, describing innovation and creation in a business environment, and continually investigating the relationship between people, experiences, and technology. Its publisher is the Association for Computing Machinery (ACM), which has its headquarters in New York City.

Steve Whittaker is a Professor in human-computer interaction at the University of California Santa Cruz. He is best known for his research at the intersection of computer science and social science in particular on computer mediated communication and personal information management. He is a Fellow of the Association for Computing Machinery (ACM), and winner of the CSCW 2018 "Lasting Impact" award. He also received a Lifetime Research Achievement Award from SIGCHI, is a Member of the SIGCHI Academy. He is Editor of the journal Human-Computer Interaction.

<span class="mw-page-title-main">Eric Horvitz</span> American computer scientist, and Technical Fellow at Microsoft

Eric Joel Horvitz is an American computer scientist, and Technical Fellow at Microsoft, where he serves as the company's first Chief Scientific Officer. He was previously the director of Microsoft Research Labs, including research centers in Redmond, WA, Cambridge, MA, New York, NY, Montreal, Canada, Cambridge, UK, and Bangalore, India.

<span class="mw-page-title-main">Jean-Daniel Fekete</span>

Jean-Daniel Fekete is a French computer scientist.

Alice Jane Bernheim Brush is an American computer scientist known for her research in human-computer interaction, ubiquitous computing and computer supported collaborative work (CSCW). She is particularly known for her research studying and building technology for homes as well as expertise conducting field studies of technology. She is the co-chair of CRA-W from 2014 to 2017.

Design fiction is a design practice aiming at exploring and criticising possible futures by creating speculative, and often provocative, scenarios narrated through designed artifacts. It is a way to facilitate and foster debates, as explained by futurist Scott Smith: "... design fiction as a communication and social object creates interactions and dialogues around futures that were missing before. It helps make it real enough for people that you can have a meaningful conversation with".

Positive computing is a technological design perspective that embraces psychological well-being and ethical practice, aiming at building a digital environment to support happier and healthier users. Positive computing develops approaches that integrate insights from psychology, education, neuroscience, and HCI with technological development. The purpose of positive computing is to bridge the technology and mental health worlds. Indeed, there are computer and mental health workshops that are aimed to bring people from both communities together.

<span class="mw-page-title-main">Projector camera systems</span>

Projector-camera systems (pro-cam), also called camera-projector systems, augment a local surface with a projected captured image of a remote surface, creating a shared workspace for remote collaboration and communication. Projector-camera systems may also be used for artistic and entertainment purposes. A pro-cam system consists of a vertical screen for implementing interpersonal space where front-facing videos are displayed, and a horizontal projected screen on the tabletop for implementing shared workspace where downward facing videos are overlapped. An automatically pre-warped image is sent to the projector to ensure that the horizontal screen appears undistorted.

Abigail Jane Sellen is a Canadian cognitive scientist, industrial engineer, and computer scientist who works for Microsoft Research in Cambridge. She is also an honorary professor at the University of Nottingham and University College London.

Animal–computer interaction (ACI) is a field of research for the design and use of technology with, for and by animals covering different kinds of animals from wildlife, zoo and domesticated animals in different roles. It emerged from, and was heavily influenced by, the discipline of Human–computer interaction (HCI). As the field expanded, it has become increasingly multi-disciplinary, incorporating techniques and research from disciplines such as artificial intelligence (AI), requirements engineering (RE), and veterinary science.

Feminist HCI is a subfield of human-computer interaction (HCI) that applies feminist theory, critical theory and philosophy to social topics in HCI, including scientific objectivity, ethical values, data collection, data interpretation, reflexivity, and unintended consequences of HCI software. The term was originally used in 2010 by Shaowen Bardzell, and although the concept and original publication are widely cited, as of 2020 Bardzell's proposed frameworks have been rarely used since.

<span class="mw-page-title-main">Wendy Mackay</span> Computer Scientist

Wendy Elizabeth Mackay is a Canadian researcher specializing in human-computer interaction. She has served in all of the roles on the SIGCHI committee, including Chair. She is a member of the CHI Academy and a recipient of a European Research Council Advanced grant. She has been a visiting professor in Stanford University between 2010 and 2012, and received the ACM SIGCHI Lifetime Service Award in 2014.

<span class="mw-page-title-main">Shumin Zhai</span> Human–computer interaction research scientist

Shumin Zhai is a Chinese-born American Canadian Human–computer interaction (HCI) research scientist and inventor. He is known for his research specifically on input devices and interaction methods, swipe-gesture-based touchscreen keyboards, eye-tracking interfaces, and models of human performance in human-computer interaction. His studies have contributed to both foundational models and understandings of HCI and practical user interface designs and flagship products. He previously worked at IBM where he invented the ShapeWriter text entry method for smartphones, which is a predecessor to the modern Swype keyboard. Dr. Zhai's publications have won the ACM UIST Lasting Impact Award and the IEEE Computer Society Best Paper Award, among others, and he is most known for his research specifically on input devices and interaction methods, swipe-gesture-based touchscreen keyboards, eye-tracking interfaces, and models of human performance in human-computer interaction. Dr. Zhai is currently a Principal Scientist at Google where he leads and directs research, design, and development of human-device input methods and haptics systems.

Tawanna Dillahunt is an American computer scientist and information scientist based at the University of Michigan School of Information. She runs the Social Innovations Group, a research group that designs, builds, and enhances technologies to solve real-world problems. Her research has been cited over 2,700 times according to Google Scholar.

Batya Friedman is an American professor in the University of Washington Information School. She is also an adjunct professor in the Paul G. Allen School Computer Science and Engineering and adjunct professor in the Department of Human-Centered Design and Engineering, where she directs the Value Sensitive Design Research Lab. She received her PhD in learning sciences from the University of California, Berkeley School of Education in 1988, and has an undergraduate degree from Berkeley in computer science and mathematics.

Jofish Kaye is an American and British scientist specializing in human-computer interaction and artificial intelligence. He runs interaction design and user research at anthem.ai, and is an editor of Personal & Ubiquitous Computing.

References

  1. Himma, Kenneth Einar; Tavani, Herman T., eds. (2008). The Handbook of Information and Computer Ethics (PDF). John Wiley & Sons Inc. ISBN   978-0-471-79959-7 . Retrieved 8 July 2016.
  2. Friedman, Batya; Hendry, David G.; Borning, Alan (2017-11-21). "A Survey of Value Sensitive Design Methods". Foundations and Trends in Human–Computer Interaction. 11 (2): 63–125. doi:10.1561/1100000015. ISSN   1551-3955. S2CID   28701004.
  3. Friedman, Batya; Kahn, Peter H.; Borning, Alan; Huldtgren, Alina (2013), Doorn, Neelke; Schuurbiers, Daan; van de Poel, Ibo; Gorman, Michael E. (eds.), "Value Sensitive Design and Information Systems", Early engagement and new technologies: Opening up the laboratory, Philosophy of Engineering and Technology, Springer Netherlands, pp. 55–95, doi:10.1007/978-94-007-7844-3_4, ISBN   9789400778443, S2CID   8176837
  4. Borning, Alan; Muller, Michael (2012). "Next steps for value sensitive design". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. CHI '12. New York, NY, USA: ACM. pp. 1125–1134. doi: 10.1145/2207676.2208560 . ISBN   9781450310154.
  5. 1 2 3 4 Friedman, Batya; Hendry, David G. (2019-05-03). Value Sensitive Design: Shaping Technology with Moral Imagination. MIT Press. ISBN   9780262351706.
  6. Friedman, Batya; Kahn, Peter H. Jr. (2002). "Value Sensitive Design: Theory and Methods". CiteSeerX   10.1.1.11.8020 .{{cite journal}}: Cite journal requires |journal= (help)
  7. Philosophy, Engineering & Technology (18 May 2010). "Value Sensitive Design: Four Challenges". slideshare.net. Archived from the original on 7 March 2016. Retrieved 24 April 2018.
  8. Umbrello, Steven (2020-04-01). "Imaginative Value Sensitive Design: Using Moral Imagination Theory to Inform Responsible Technology Design". Science and Engineering Ethics. 26 (2): 575–595. doi:10.1007/s11948-019-00104-4. hdl: 2318/1699361 . ISSN   1471-5546. PMID   30972629. S2CID   108295110.
  9. van den Hoven, Jeroen (2007). "ICT and Value Sensitive Design". In Goujon, Philippe; Lavelle, Sylvian; Duquenoy, Penny; Kimppa, Kai; Laurent, Véronique (eds.). The Information Society: Innovation, Legitimacy, Ethics and Democracy in honor of Professor Jacques Berleur s.j. IFIP International Federation for Information Processing. Vol. 233. Boston, MA: Springer US. pp. 67–72. doi: 10.1007/978-0-387-72381-5_8 . ISBN   978-0-387-72381-5.
  10. 1 2 Longo, Francesco; Padovano, Antonio; Umbrello, Steven (January 2020). "Value-Oriented and Ethical Technology Engineering in Industry 5.0: A Human-Centric Perspective for the Design of the Factory of the Future". Applied Sciences. 10 (12): 4182. doi: 10.3390/app10124182 . hdl: 2318/1741791 .
  11. Spiekermann, Sarah (July 2012). "The challenges of privacy by design". Communications of the ACM. 55 (7): 38–40. doi:10.1145/2209249.2209263. ISSN   0001-0782. S2CID   3023111.
  12. van Wynsberghe, Aimee (2013-06-01). "Designing Robots for Care: Care Centered Value-Sensitive Design". Science and Engineering Ethics. 19 (2): 407–433. doi:10.1007/s11948-011-9343-6. ISSN   1471-5546. PMC   3662860 . PMID   22212357.
  13. Friedman, B., Kahn Jr, P. H., Borning, A., & Kahn, P. H. (2006). Value Sensitive Design and information systems. Human-Computer Interaction and Management Information Systems: Foundations. ME Sharpe, New York, 348–372.
  14. 1 2 Friedman, Batya; Kahn, Peter H. Jr.; Hagman, Jennifer; Severson, Rachel L.; Gill, Brian (2006-05-01). "The Watcher and the Watched: Social Judgments About Privacy in a Public Place". Human–Computer Interaction. 21 (2): 235–272. doi:10.1207/s15327051hci2102_3. ISSN   0737-0024. S2CID   54165089.
  15. 1 2 Nathan, Lisa P.; Friedman, Batya; Klasnja, Predrag; Kane, Shaun K.; Miller, Jessica K. (2008-02-25). "Envisioning systemic effects on persons and society throughout interactive system design". Proceedings of the 7th ACM conference on Designing interactive systems. DIS '08. Cape Town, South Africa: Association for Computing Machinery. pp. 1–10. doi:10.1145/1394445.1394446. ISBN   978-1-60558-002-9. S2CID   2412766.
  16. 1 2 3 4 5 Czeskis, Alexei; Dermendjieva, Ivayla; Yapit, Hussein; Borning, Alan; Friedman, Batya; Gill, Brian; Kohno, Tadayoshi (2010-07-14). "Parenting from the pocket". Proceedings of the Sixth Symposium on Usable Privacy and Security. SOUPS '10. Redmond, Washington, USA: Association for Computing Machinery. pp. 1–15. doi:10.1145/1837110.1837130. ISBN   978-1-4503-0264-7. S2CID   13951473.
  17. Watkins, Kari Edison; Ferris, Brian; Malinovskiy, Yegor; Borning, Alan (2013-11-27). "Beyond Context-Sensitive Solutions: Using Value-Sensitive Design to Identify Needed Transit Information Tools". Urban Public Transportation Systems 2013. pp. 296–308. doi:10.1061/9780784413210.026. ISBN   9780784413210.
  18. Yoo, Daisy (2018-08-04). "Stakeholder Tokens: a constructive method for value sensitive design stakeholder analysis". Ethics and Information Technology. 23: 63–67. doi:10.1007/s10676-018-9474-4. ISSN   1572-8439. S2CID   52048390.
  19. 1 2 Borning, Alan; Friedman, Batya; Davis, Janet; Lin, Peyina (2005). Gellersen, Hans; Schmidt, Kjeld; Beaudouin-Lafon, Michel; Mackay, Wendy (eds.). "Informing Public Deliberation: Value Sensitive Design of Indicators for a Large-Scale Urban Simulation". Ecscw 2005. Dordrecht: Springer Netherlands: 449–468. doi:10.1007/1-4020-4023-7_23. ISBN   978-1-4020-4023-8. S2CID   17369120.
  20. Friedman, Batya; Smith, Ian; H. Kahn, Peter; Consolvo, Sunny; Selawski, Jaina (2006). "Development of a Privacy Addendum for Open Source Licenses: Value Sensitive Design in Industry". In Dourish, Paul; Friday, Adrian (eds.). UbiComp 2006: Ubiquitous Computing. Lecture Notes in Computer Science. Vol. 4206. Berlin, Heidelberg: Springer. pp. 194–211. doi:10.1007/11853565_12. ISBN   978-3-540-39635-2.
  21. 1 2 3 Miller, Jessica K.; Friedman, Batya; Jancke, Gavin; Gill, Brian (2007-11-04). "Value tensions in design". Proceedings of the 2007 international ACM conference on Conference on supporting group work - GROUP '07. Sanibel Island, Florida, USA: Association for Computing Machinery. pp. 281–290. doi:10.1145/1316624.1316668. ISBN   978-1-59593-845-9. S2CID   2633485.
  22. Friedman, Batya; Hurley, David; Howe, Daniel C.; Felten, Edward; Nissenbaum, Helen (2002-04-20). "Users' conceptions of web security: A comparative study". CHI '02 Extended Abstracts on Human Factors in Computing Systems. CHI EA '02. Minneapolis, Minnesota, USA: Association for Computing Machinery. pp. 746–747. doi:10.1145/506443.506577. ISBN   978-1-58113-454-4. S2CID   27784060.
  23. Woelfer, Jill Palzkill; Iverson, Amy; Hendry, David G.; Friedman, Batya; Gill, Brian T. (2011-05-07). "Improving the safety of homeless young people with mobile phones". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. CHI '11. Vancouver, BC, Canada: Association for Computing Machinery. pp. 1707–1716. doi:10.1145/1978942.1979191. ISBN   978-1-4503-0228-9. S2CID   41591259.
  24. 1 2 Friedman, Batya (1997-08-01). "Social Judgments and technological innovation: Adolescents' understanding of property, privacy, and electronic information". Computers in Human Behavior. 13 (3): 327–351. doi:10.1016/S0747-5632(97)00013-7. ISSN   0747-5632.
  25. 1 2 Freier, Nathan G. (2008-04-06). "Children attribute moral standing to a personified agent". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. CHI '08. Florence, Italy: Association for Computing Machinery. pp. 343–352. doi:10.1145/1357054.1357113. ISBN   978-1-60558-011-1. S2CID   15580819.
  26. Umbrello, Steven; van de Poel, Ibo (2021). "Mapping value sensitive design onto AI for social good principles". AI and Ethics. Springer Nature. 1 (3): 283–296. doi:10.1007/s43681-021-00038-3. PMC   7848675 . PMID   34790942. S2CID   231744217.
  27. Munson, Sean A.; Avrahami, Daniel; Consolvo, Sunny; Fogarty, James; Friedman, Batya; Smith, Ian (2011-06-12). "Attitudes toward online availability of US public records". Proceedings of the 12th Annual International Digital Government Research Conference: Digital Government Innovation in Challenging Times. dg.o '11. College Park, Maryland, USA: Association for Computing Machinery. pp. 2–9. doi:10.1145/2037556.2037558. ISBN   978-1-4503-0762-8. S2CID   10276344.
  28. Woelfer, Jill Palzkill; Hendry, David G. (2009). "Stabilizing homeless young people with information and place". Journal of the American Society for Information Science and Technology. 60 (11): 2300–2312. doi:10.1002/asi.21146. ISSN   1532-2890. S2CID   18725449.
  29. 1 2 Denning, Tamara; Borning, Alan; Friedman, Batya; Gill, Brian T.; Kohno, Tadayoshi; Maisel, William H. (2010-04-10). "Patients, pacemakers, and implantable defibrillators". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. CHI '10. Atlanta, Georgia, USA: Association for Computing Machinery. pp. 917–926. doi:10.1145/1753326.1753462. ISBN   978-1-60558-929-9. S2CID   16571765.
  30. 1 2 3 Yoo, Daisy; Huldtgren, Alina; Woelfer, Jill Palzkill; Hendry, David G.; Friedman, Batya (2013-04-27). "A value sensitive action-reflection model". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. CHI '13. Paris, France: Association for Computing Machinery. pp. 419–428. doi:10.1145/2470654.2470715. ISBN   978-1-4503-1899-0. S2CID   2603883.
  31. Nathan, Lisa P. (2012). "Sustainable information practice: An ethnographic investigation". Journal of the American Society for Information Science and Technology. 63 (11): 2254–2268. doi:10.1002/asi.22726. ISSN   1532-2890.
  32. Millett, Lynette I.; Friedman, Batya; Felten, Edward (2001-03-01). "Cookies and Web browser design". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. CHI '01. Seattle, Washington, USA: Association for Computing Machinery. pp. 46–52. doi:10.1145/365024.365034. ISBN   978-1-58113-327-1. S2CID   1596706.
  33. 1 2 Yoo, Daisy; Derthick, Katie; Ghassemian, Shaghayegh; Hakizimana, Jean; Gill, Brian; Friedman, Batya (2016-05-07). "Multi-lifespan Design Thinking". Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems. CHI '16. San Jose, California, USA: Association for Computing Machinery. pp. 4423–4434. doi: 10.1145/2858036.2858366 . ISBN   978-1-4503-3362-7. S2CID   2148594.
  34. Kaptein, Maurits; Eckles, Dean; Davis, Janet (2011-09-01). "Envisioning persuasion profiles: challenges for public policy and ethical practice". Interactions. 18 (5): 66–69. doi:10.1145/2008176.2008191. ISSN   1072-5520. S2CID   11099713.
  35. Friedman, Batya; Hendry, David (2012-05-05). "The envisioning cards". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. CHI '12. Austin, Texas, USA: Association for Computing Machinery. pp. 1145–1148. doi:10.1145/2207676.2208562. ISBN   978-1-4503-1015-4. S2CID   24059203.
  36. 1 2 Umbrello, Steven (2020-10-30). "Combinatory and Complementary Practices of Values and Virtues in Design: A Reply to Reijers and Gordijn". Filosofia (in Italian) (65): 107–121 Paginazione. doi:10.13135/2704-8195/5236.
  37. 1 2 Reijers, Wessel; Gordijn, Bert (2019-05-13). "Moving from value sensitive design to virtuous practice design". Journal of Information, Communication and Ethics in Society. 17 (2): 196–209. doi:10.1108/JICES-10-2018-0080. hdl: 1814/63270 . ISSN   1477-996X. S2CID   197695970.
  38. 1 2 Le Dantec, Christopher A.; Poole, Erika Shehan; Wyche, Susan P. (2009). "Values as lived experience". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. New York, New York, USA: ACM Press. p. 1141. doi:10.1145/1518701.1518875. ISBN   978-1-60558-246-7. S2CID   13933217.
  39. 1 2 Manders-Huits, Noëmi (2011-06-01). "What Values in Design? The Challenge of Incorporating Moral Values into Design". Science and Engineering Ethics. 17 (2): 271–287. doi:10.1007/s11948-010-9198-2. ISSN   1471-5546. PMC   3124645 . PMID   20224927.
  40. Umbrello, Steven (2018-05-04). "The moral psychology of value sensitive design: the methodological issues of moral intuitions for responsible innovation". Journal of Responsible Innovation. 5 (2): 186–200. doi:10.1080/23299460.2018.1457401. hdl: 2318/1685524 . ISSN   2329-9460.
  41. van de Poel, Ibo (September 2020). "Embedding Values in Artificial Intelligence (AI) Systems". Minds and Machines. 30 (3): 385–409. doi: 10.1007/s11023-020-09537-4 . ISSN   0924-6495. S2CID   222354603.
  42. Floridi, Luciano; Cowls, Josh; King, Thomas C.; Taddeo, Mariarosaria (June 2020). "How to Design AI for Social Good: Seven Essential Factors". Science and Engineering Ethics. 26 (3): 1771–1796. doi:10.1007/s11948-020-00213-5. ISSN   1353-3452. PMC   7286860 . PMID   32246245.
  43. Umbrello, Steven; van de Poel, Ibo (2021-02-01). "Mapping value sensitive design onto AI for social good principles". AI and Ethics. 1 (3): 283–296. doi:10.1007/s43681-021-00038-3. ISSN   2730-5953. PMC   7848675 . PMID   34790942.