Social information seeking

Last updated

Social information seeking is a field of research that involves studying situations, motivations, and methods for people seeking and sharing information in participatory online social sites, such as Yahoo! Answers, Answerbag, WikiAnswers and Twitter as well as building systems for supporting such activities. Highly related topics involve traditional and virtual reference services, information retrieval, information extraction, and knowledge representation. [1]

Contents

Background

Social information seeking is often materialized in online question-answering (QA) websites, which are driven by a community. Such QA sites have emerged in the past few years as an enormous market, so to speak, for the fulfillment of information needs. Estimates of the volume of questions answered are difficult to come by, but it is likely that the number of questions answered on social/community QA (cQA) sites far exceeds the number of questions answered by library reference services, [2] which until recently were one of the few institutional sources for such question answering. cQA sites make their content – questions and associated answers submitted on the site – available on the open web, and indexable by search engines, thus enabling web users to find answers provided for previously asked questions in response to new queries.

The popularity of such sites have been increasing dramatically for the past several years. Major sites that provide a general platform for questions of all types include Yahoo! Answers, Answerbag and Quora. While other sites that focus on particular fields; for example, StackOverflow (computing). StackOverflow has 3.45 million questions, 1.3 million users and over 6.86 million answers since July 2008 while Quora has 437 thousand questions, 264 thousand users and 979 thousand answers. [3]

Social Q&A or cQA, according to Shah et al., [4] consists of three components: a mechanism for users to submit questions in natural language, a venue for users to submit answers to questions, and a community built around this exchange. Viewed in that light, online communities have performed a question answering function perhaps since the advent of Usenet and Bulletin Board Systems, so in one sense cQA is nothing new. Websites dedicated to cQA, however, have emerged on the web only within the past few years: the first cQA site was the Korean Naver Knowledge iN, launched in 2002, while the first English-language CQA site was Answerbag, launched in April 2003. Despite this short history, however, cQA has already attracted a great deal of attention from researchers investigating information seeking behaviors, [5] selection of resources, [6] social annotations, [7] user motivations, [8] comparisons with other types of question answering services, [9] and a range of other information-related behaviors.

Research questions

Some of the interesting and important research questions in this area include:

Shah et al. [10] provide a detailed research agenda for social Q&A. A new book by Shah [11] presents a more recent and comprehensive information pertaining to social information seeking.

Friendsourcing in social Q&A

Friendsourcing is an important component of social question and answering, including how to route questions to friends or others who will most likely answer the question. [12] The important questions include what people's behaviors are in social networks, especially what kinds of questions people ask from their social networks and how different question types affect the frequency, speed and quality of answers they receive.

Morris et al. (2010) [13] conducted a survey of question and answering within social networks with 624 people, and gathered detailed data about the behavior of Q&A, including frequency, types of questions and answers, and motivations. They found that half (50.6%) of respondents reported having used their status messages to ask a question, which indicated that Q&A on social networks is popular. Also, the types of questions people asked include recommendation, opinion, factual knowledge, rhetorical, etc. And motivations for asking include trust, asking subjective questions, etc. Their analysis also explored the relationships between answer speed and quality, questions’ property and participants’ property. Only a very small portion (6.5%) of the questions were answered, but the 89.3% of the respondents were satisfied with the response time they experienced even though there's a discrepancy between that and expectation. Also, the responses gathered via social networks appear to be very valuable. Their findings implied design for search tools that could combine the speed and breadth of traditional search engines with the trustworthiness, personalization, and the high engagement of social media Q&A.

Paul et al. (2011) [14] did a study on question and answering on Twitter, and found that out of the 1152 questions they examined, the most popular question types asked on Twitter were rhetorical (42%) and factual (16%). Surprisingly, along with entertainment (29%) and technology (29%) questions, people asked personal and health-related questions (11%). Only 18.7% questions received response, while a handful of questions received a high number of responses. The larger the askers’ network, the more responses she received; however, posting more tweets or posting more frequently did not increase chances of receiving a response. Most often the “follow” relationship between asker and answerer was one-way. Paul et al. also examined what factors of the askers would increase the chance of getting a response and found that more relevant responses are received when there is a mutual relationship between askers and answerers. Intuitively, we would expect this, as mutual relationship would indicate stronger tie strength and hence, more number of relevant answers.

Social Q&A services

Existing social Q&A services can be characterized from the three perspectives, by the definition of social Q&A as a service involving (1) a method for presenting information needs, (2) a place for responding to information need, and (3) participation as a community.

Presenting information needs

These social networks support various friendsourcing behavior, provide information benefits that oftentimes traditional search tools cannot, and also may reinforce social bonds through the process. However, there are many questions and limitations that may prevent people from asking questions on their social networks. For example, they may feel uncomfortable asking questions that are too private, might not want to cost too much other people's time and effort, or might feel the burden of social debts.

Rzeszotarski and Morris (2014) [15] took a novel approach to explore the perceived social costs of friendsourcing on Twitter via monetary choices. They modeled friendsourcing costs across users, and compared it with crowdsourcing on Amazon Mechanical Turk. Their findings suggested interesting design considerations for minimizing social cost by building a hybrid system combining friendsourcing and crowdsourcing with microtask markets.

Responding to information needs

Sometimes, only asking question from people's own social networks or friends is not enough. If the question is obscure or time sensitive, no members of their social networks may know the answer. For example, this person's friends might not have expertise in providing evaluations for a specific model of digital camera. Also asking the current wait time for security at the local airport might not be possible if none of this person's friends are currently at the airport.

Nichols and Kang (2012) [16] leveraged Twitter for question and answering with targeted strangers by taking advantage of its public accessibility. In their approach, they mined the public status updates posted on Twitter to find strangers with potentially useful information, and send questions to these strangers to collect responses. As a feasibility study, they collected information regarding response rate, and response time. 42% of users responded to questions from strangers, and 44% of the responses arrived within 30 minutes.

Participation as a community

Another important and unique component of social Q&A system is that it is a community which allows members to form relationships and bonds, so that their behavior in these social Q&A services will also add to their social capital.

Gray et al. (2013) [17] explored how bridging social capital, question type and relational closeness influence the perceived usefulness and satisfaction of information obtained through questions asked on Facebook. Their results indicated that bridging social capital could positively predict the perceived utility of the acquired information, meaning that information exchanges on social networks is an effective way of social capital conversion. Also, useful answers are more likely to be received from weak ties than strong ties.

Authority detection in social media

In order to recommend the most appropriate users to provide answers in a social network, we need to find approaches to detect users' authority in a social network. In the field of information retrieval, there has been a trend of research investigating ways to detect users' authority effectively and accurately in a social network.

Cha et al. [18] investigate possible metrics for determining authority users on popular social network Twitter. They propose the following three simple network-based metrics and discuss their usefulness in determining a user's influence.

  1. indegree (followers count)
  2. retweet count
  3. mention count

An initial analysis of the three aforementioned metrics showed that the users with the highest indegrees and the users with the highest retweet/mention counts were not the same. The top 1% of users by indegree are shown to have very low correlation with the same percentile of users by retweets and by mentions. This implies that follower count is not useful in determining whether a user's tweets get retweeted or whether the other users engage with them.

Pal et al. [19] designed features to measure a user's authority on a certain topic. For example, retweet impact refers to how many times a certain user has been retweeted on a certain topic. The impact is dampened by a factor measuring how many times the user had been retweeted by a unique author to avoid the cases when a user has fans who retweet regardless of the content. They first used a clustering approach to find the target cluster which has the highest average score across all features, and used a ranking algorithm to find the most authoritative users within the cluster.

With these authority detection methods, social Q&A could be more effective in providing accurate answers to askers.

Related Research Articles

Question answering (QA) is a computer science discipline within the fields of information retrieval and natural language processing (NLP), which is concerned with building systems that automatically answer questions posed by humans in a natural language.

Opinion leadership is leadership by an active media user who interprets the meaning of media messages or content for lower-end media users. Typically opinion leaders are held in high esteem by those who accept their opinions. Opinion leadership comes from the theory of two-step flow of communication propounded by Paul Lazarsfeld and Elihu Katz. Significant developers of the opinion leader concept have been Robert K. Merton, C. Wright Mills and Bernard Berelson. This theory is one of several models that try to explain the diffusion of innovations, ideas, or commercial products.

In Internet culture, a lurker is typically a member of an online community who observes, but does not participate. The exact definition depends on context. Lurkers make up a large proportion of all users in online communities. Lurking allows users to learn the conventions of an online community before they participate, improving their socialization when they eventually "de-lurk". However, a lack of social contact while lurking sometimes causes loneliness or apathy among lurkers.

Yahoo! Answers was a community-driven question-and-answer (Q&A) website or knowledge market owned by Yahoo! where users would ask questions and answer those submitted by others, and upvote them to increase their visibility. Questions were organised into categories with multiple sub-categories under each to cover every topic users may ask questions on, such as beauty, business, finance, cars, electronics, entertainment, games, gardening, science, news, politics, parenting, pregnancy, and travel. The number of poorly formed questions and inaccurate answers made the site a target of ridicule.

Exploratory search is a specialization of information exploration which represents the activities carried out by searchers who are:

Social search is a behavior of retrieving and searching on a social searching engine that mainly searches user-generated content such as news, videos and images related search queries on social media like Facebook, LinkedIn, Twitter, Instagram and Flickr. It is an enhanced version of web search that combines traditional algorithms. The idea behind social search is that instead of ranking search results purely based on semantic relevance between a query and the results, a social search system also takes into account social relationships between the results and the searcher. The social relationships could be in various forms. For example, in LinkedIn people search engine, the social relationships include social connections between searcher and each result, whether or not they are in the same industries, work for the same companies, belong the same social groups, and go the same schools, etc.

Social information processing is "an activity through which collective human actions organize knowledge." It is the creation and processing of information by a group of people. As an academic field Social Information Processing studies the information processing power of networked social systems.

<span class="mw-page-title-main">Stack Overflow</span> Website hosting questions and answers on a wide range of topics in computer programming

Stack Overflow is a question and answer website for professional and enthusiast programmers. It is the flagship site of the Stack Exchange Network. It was created in 2008 by Jeff Atwood and Joel Spolsky. It features questions and answers on a wide range of topics in computer programming. It was created to be a more open alternative to earlier question and answer websites such as Experts-Exchange. Stack Overflow was sold to Prosus, a Netherlands-based consumer internet conglomerate, on 2 June 2021 for $1.8 billion.

Online participation is used to describe the interaction between users and online communities on the web. Online communities often involve members to provide content to the website and/or contribute in some way. Examples of such include wikis, blogs, online multiplayer games, and other types of social platforms. Online participation is currently a heavily researched field. It provides insight into fields such as web design, online marketing, crowdsourcing, and many areas of psychology. Some subcategories that fall under online participation are: commitment to online communities, coordination & interaction, and member recruitment.

Collaborative information seeking (CIS) is a field of research that involves studying situations, motivations, and methods for people working in collaborative groups for information seeking projects, as well as building systems for supporting such activities. Such projects often involve information searching or information retrieval (IR), information gathering, and information sharing. Beyond that, CIS can extend to collaborative information synthesis and collaborative sense-making.

MathOverflow is a mathematics question-and-answer (Q&A) website, which serves as an online community of mathematicians. It allows users to ask questions, submit answers, and rate both, all while getting merit points for their activities. It is a part of the Stack Exchange Network.

<span class="mw-page-title-main">Quora</span> Question-and-answer platform

Quora is a social question-and-answer website based in Mountain View, California. It was founded on June 25, 2009, and made available to the public on June 21, 2010. Users can collaborate by editing questions and commenting on answers that have been submitted by other users. As of 2020, the website was visited by 300 million users a month.

<span class="mw-page-title-main">Blurtit</span> British Q&A website

Blurtit is a British Q&A website where people asked questions and a community of regular users provided answers based on their knowledge or opinions. Blurtit was founded in 2006, and was based in Norwich in Norfolk, UK.

Reblogging is the mechanism in blogging which allows users to repost the content of another user's post with an indication that the source of the post is another user.

Ed Huai-Hsin Chi is a Taiwanese American computer scientist and research scientist at Google, known for his early work in applying the theory of information scent to predict usability of websites.

Adam N. Joinson is a British author, academic and public speaker within the area of cyberpsychology. He is Professor of Information Systems at University of Bath, following posts at the University of West of England and the Open University. and has conducted ground breaking research into the psychology of Internet usage.

Social Visualization is an interdisciplinary intersection of information visualization to study creating intuitive depictions of massive and complex social interactions for social purposes. By visualizing those interactions made not only in the cyberspace including social media but also the physical world, captured through sensors, it can reveal overall patterns of social memes or it highlights one individual’s implicit behaviors in diverse social spaces. In particular, it is the study “primarily concerned with the visualization of text, audio, and visual interaction data to uncover social connections and interaction patterns in online and physical spaces. ACM Computing Classification System has classified this field of study under the category of Human-Centered Computing (1st) and Information Visualization (2nd) as a third level concept in a general sense.

Spring.me was a social networking service. Until a rebranding in 2013, it was known as Formspring, a question-and-answer-based social network launched in 2009 by Ade Olonoh, the founder of online form builder Formstack.

Q&A software is online software that attempts to answer questions asked by users. Q&A software is frequently integrated by large and specialist corporations and tends to be implemented as a community that allows users in similar fields to discuss questions and provide answers to common and specialist questions.

Social media mining is the process of obtaining big data from user-generated content on social media sites and mobile apps in order to extract actionable patterns, form conclusions about users, and act upon the information, often for the purpose of advertising to users or conducting research. The term is an analogy to the resource extraction process of mining for rare minerals. Resource extraction mining requires mining companies to shift through vast quantities of raw ore to find the precious minerals; likewise, social media mining requires human data analysts and automated software programs to shift through massive amounts of raw social media data in order to discern patterns and trends relating to social media usage, online behaviours, sharing of content, connections between individuals, online buying behaviour, and more. These patterns and trends are of interest to companies, governments and not-for-profit organizations, as these organizations can use these patterns and trends to design their strategies or introduce new programs, new products, processes or services.

References

  1. Čižmešija, Antonela (March 2018). "Students' Information Seeking Behavior in Online Environment Using Web 2.0 Tools". INTED2018 Proceedings. IATED. doi:10.21125/inted.2018.1636. ISBN   978-84-697-9480-7.
  2. Janes, J. (2003). The Global Census of Digital Reference. In 5th Annual VRD Conference. San Antonio, TX.
  3. Wang, G., Gill, K., Mohanlal, M., Zheng, H., & Zhao, B. Y. (2013, May). Wisdom in the social crowd: an analysis of Quora. In Proceedings of the 22nd international conference on World Wide Web (pp. 1341-1352). International World Wide Web Conferences Steering Committee
  4. Shah, C., Oh, S., & Oh, J-S. (2009). Research Agenda for Social Q&A. Library and Information Science Research, 11(4), 205-209.
  5. Kim, S., Oh, J-S., & Oh, S. (2007). Best-Answer Selection Criteria in a Social Q&A site from the User Oriented Relevance Perspective. Proceedings of the 70th Annual Meeting of the American Society for Information Science and Technology (ASIST ‘07), 44.
  6. Harper, M. F., Raban, D. R., Rafaeli, S., & Konstan, J. K. (2008). Predictors of answer quality in online Q&A sites. In Proceedings of the 26th Annual SIGCHI Conference on Human Factors in Computing Systems (pp. 865−874). New York: ACM.
  7. Gazan, R. (2008). Social annotations in digital library collections. D-Lib Magazine, 11/12(14). Available from http://www.dlib.org/dlib/november08/gazan/11gazan.html.
  8. Shah, C., Oh, J. S., & Oh, S. (2008). Exploring characteristics and effects of user participation in online social Q&A sites. First Monday, 13(9). Available from http://www.uic.edu/htbin/cgiwrap/bin/ojs/index.php/fm/article/view/2182/2028.
  9. Su, Q., Pavlov, D., Chow, J., & Baker, W. (2007). Internet-scale collection of human- reviewed data. In C. L. Williamson, M. E. Zurko, P. E. Patel-Schneider, & P. J. Shenoy (Eds.), Proceedings of the 16th International Conference on World Wide Web (pp. 231−240). New York: ACM.
  10. Shah, C., Oh, S., & Oh, J. S. (2009). Research agenda for social Q&A. Library & Information Science Research, 31(4), 205-209. Retrieved January 2, 2011.
  11. Shah, C. (2017). Social Information Seeking: Leveraging the Wisdom of the Crowd. The Information Retrieval (IR) series. Berlin, Germany: Springer. ISBN   978-3-319-56756-3. (187 pages)
  12. Liu, Z., and Jansen, B. J. (2015) Analysis of Question and Answering Behavior in Question Routing Services. 21st International Conference on Collaboration and Technology (CRIWG 2015). Yerevan, Armenia. 22 – 25 Sept., p.72-85
  13. Morris, Meredith Ringel; Teevan, Jaime; Panovich, Katrina (2010-01-01). "What Do People Ask Their Social Networks, and Why?: A Survey Study of Status Message Q&a Behavior". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. CHI '10. New York, NY, USA: ACM: 1739–1748. doi:10.1145/1753326.1753587. ISBN   9781605589299.
  14. Paul, S. A., Hong, L., & Chi, E. H. (2011, May). Is Twitter a Good Place for Asking Questions? A Characterization Study. In ICWSM.
  15. Rzeszotarski, Jeffrey M.; Morris, Meredith Ringel (2014-01-01). "Estimating the Social Costs of Friendsourcing". Proceedings of the 32nd Annual ACM Conference on Human Factors in Computing Systems. CHI '14. New York, NY, USA: ACM: 2735–2744. doi:10.1145/2556288.2557181. ISBN   9781450324731.
  16. Nichols, Jeffrey; Kang, Jeon-Hyung (2012-01-01). "Asking Questions of Targeted Strangers on Social Networks". Proceedings of the ACM 2012 Conference on Computer Supported Cooperative Work. CSCW '12. New York, NY, USA: ACM: 999–1002. doi:10.1145/2145204.2145352. ISBN   9781450310864.
  17. Gray, Rebecca; Ellison, Nicole B.; Vitak, Jessica; Lampe, Cliff (2013-01-01). "Who Wants to Know?: Question-asking and Answering Practices Among Facebook Users". Proceedings of the 2013 Conference on Computer Supported Cooperative Work. CSCW '13. New York, NY, USA: ACM: 1213–1224. doi:10.1145/2441776.2441913. ISBN   9781450313315.
  18. Cha, M., Haddadi, H., Benevenuto, F., & Gummadi, P. K. (2010). Measuring User Influence in Twitter: The Million Follower Fallacy. ICWSM, 10(10-17), 30.
  19. Pal, A., & Counts, S. (2011, February). Identifying topical authorities in microblogs. In Proceedings of the fourth ACM international conference on Web search and data mining (pp. 45-54). ACM.

People associated with social information seeking include: