Discoverability

Last updated

Discoverability is the degree to which something, especially a piece of content or information, can be found in a search of a file, database, or other information system. Discoverability is a concern in library and information science, many aspects of digital media, software and web development, and in marketing, since products and services cannot be used if people cannot find it or do not understand what it can be used for.

Contents

Metadata, or "information about information," such as a book's title, a product's description, or a website's keywords, affects how discoverable something is on a database or online. Adding metadata to a product that is available online can make it easier for end users to find the product. For example, if a song file is made available online, making the title, name of the band, genre, year of release, and other pertinent information available in connection with this song means the file can be retrieved more easily. Organizing information by putting it into alphabetical order or including it in a search engine is an example of how to improve discoverability.

Discoverability is related to, but different from, accessibility and usability, other qualities that affect the usefulness of a piece of information.

Etymology

The concept of "discoverability" in an information science and online context is a loose borrowing from the concept of the similar name in the legal profession. In law, "discovery" is a pre-trial procedure in a lawsuit in which each party, through the law of civil procedure, can obtain evidence from the other party or parties by means of discovery devices such as a request for answers to interrogatories, request for production of documents, request for admissions and depositions. Discovery can be obtained from non-parties using subpoenas. When a discovery request is objected to, the requesting party may seek the assistance of the court by filing a motion to compel discovery. [1]

Purpose

The usability of any piece of information directly relates to how discoverable it is, either in a "walled garden" database or on the open Internet. The quality of information available on this database or on the Internet depends upon the quality of the meta-information about each item, product, or service. In the case of a service, because of the emphasis placed on service reusability, opportunities should exist for reuse of this service. However, reuse is only possible if information is discoverable in the first place. To make items, products, and services discoverable, the process is as follows:

  1. Document the information about the item, product or service (the metadata) in a consistent manner.
  2. Store the documented information (metadata) in a searchable repository.
    • while technically a human-searchable repository, such as a printed paper list would qualify, "searchable repository" is usually taken to mean a computer-searchable repository, such as a database that a human user can search using some type of search engine or "find" feature.
  3. Enable search for the documented information in an efficient manner.
    • supports number 2, because while reading through a printed paper list by hand might be feasible in a theoretical sense, it is not time and cost-efficient in comparison with computer-based searching.

Apart from increasing the reuse potential of the services, discoverability is also required to avoid development of solution logic that is already contained in an existing service. To design services that are not only discoverable but also provide interpretable information about their capabilities, the service discoverability principle provides guidelines that could be applied during the service-oriented analysis phase of the service delivery process.

Specific to digital media

In relation to audiovisual content, according to the meaning given by the Canadian Radio-television and Telecommunications Commission (CRTC) for the purpose of its 2016 Discoverability Summit, discoverability can be summed up to the intrinsic ability of given content to "stand out of the lot", or to position itself so as to be easily found and discovered. [2] A piece of audiovisual content can be a movie, a TV series, music, a book (eBook), an audio book or podcast. When audiovisual content such as a digital file for a TV show, movie, or song, is made available online, if the content is "tagged" with identifying information such as the names of the key artists (e.g., actors, directors and screenwriters for TV shows and movies; singers, musicians and record producers for songs) and the genres (for movies genres, music genres, etc.).

When users interact with online content, algorithms typically determine what types of content the user is interested in, and then a computer program suggests "more like this", which is other content that the user may be interested in. Different websites and systems have different algorithms, but one approach, used by Amazon (company) for its online store, is to indicate to a user: "customers who bought x also bought y" (affinity analysis, collaborative filtering). This example is oriented around online purchasing behaviour, but an algorithm could also be programmed to provide suggestions based on other factors (e.g., searching, viewing, etc.).[ citation needed ]

Discoverability is typically referred to in connection with search engines. A highly "discoverable" piece of content would appear at the top, or near the top of a user's search results. A related concept is the role of "recommendation engines", which give a user recommendations based on his/her previous online activity. Discoverability applies to computers and devices that can access the Internet, including various console video game systems and mobile devices such as tablets and smartphones. When producers make an effort to promote content (e.g., a TV show, film, song, or video game), they can use traditional marketing (billboards, TV ads, radio ads) and digital ads (pop-up ads, pre-roll ads, etc.), or a mix of traditional and digital marketing.[ citation needed ]

Even before the user's intervention by searching for a certain content or type of content, discoverability is the prime factor which contributes to whether a piece of audiovisual content will be likely to be found in the various digital modes of content consumption. As of 2017, modes of searching include looking on Netflix for movies, Spotify for music, Audible for audio books, etc., although the concept can also more generally be applied to content found on Twitter, Tumblr, Instagram, and other websites. It involves more than a content's mere presence on a given platform; it can involve associating this content with "keywords" (tags), search algorithms, positioning within different categories, metadata, etc. Thus, discoverability enables as much as it promotes. For audiovisual content broadcast or streamed on digital media using the Internet, discoverability includes the underlying concepts of information science and programming architecture, which are at the very foundation of the search for a specific product, information or content.

Applications

Within a webpage

Within a specific webpage or software application ("app"), the discoverability of a feature, content or link depends on a range of factors, including the size, colour, highlighting features, and position within the page. When colour is used to communicate the importance of a feature or link, designers typically use other elements as well, such as shadows or bolding, for individuals, who cannot see certain colours. Just as traditional paper printing created other physical locations that stood out, such as being "above the fold" of a newspaper versus "below the fold", a web page or app's screenview may have certain locations that give features additional visibility to users, such as being right at the bottom of the web page or screen.[ citation needed ]

The positional advantages or disadvantages of various locations depend on different cultures and languages (e.g., left to right vs. right to left). Some locations have become established, such as having toolbars at the top of a screen or webpage. Some designers have argued that commonly used features (e.g., a print button) should be much more visually prominent than very rarely used features. Some features cannot be seen, but there is a convention that if the user places the mouse cursor in a certain area, then a toolbar or function option will become visible. In general, because of the smaller screen of mobile devices, controls are often not placed right in the centre of the screen, because that is where the user views content or text.

Some organizations try to increase the discoverability of a certain feature by adding animation, such as a moving "click here" icon. As of 2017, the addition of motion sensors and geotracking to mobile devices has made webpage design for discoverability more complex, because smartphones and tablets are typically capable of having many more inputs from the user than a 1980s era desktop, including "swiping" the touchscreen, touching images on the screen, or tilting the device. One of the challenges in webpage and app design is that the degree of sophistication and experience of users with navigating in the webpage or app environment varies a great deal, from individuals who are new to using these applications at one extreme to experienced computer users.

For items that are searched for online, the goal of discoverability is to be at or near the top of the search results. Organizations may make efforts to make it more likely, that "their" content or webpages are at the top, or close to the top, of search results; these approaches are often collectively called search engine optimization (SEO). Note that when an organization takes action to increase the SEO of its website, this does not normally involve changes to the search engine itself; rather, it involves adding metadata tags and original content, among other strategies, to increase the "visibility" of the website to search engine algorithms. [3]

Services

In a service delivery context, the application of this principle requires collecting information about the service during the service analysis phase as during this phase; maximum information is available about the service's functional context [4] and the capabilities of the service. At this stage, the domain knowledge of the business experts could also be enlisted to document meta-data about the service. In the service-oriented design phase, the already gathered meta-data could be made part of the service contract. [5] The OASIS SOA-RM standard specifies service description as an artifact that represents service meta-data. [6]

To make the service meta-data accessible to interested parties, it must be centrally accessible. This could either be done by publishing the service-meta to a dedicated 'service registry' [7] or by simply placing this information in a 'shared directory'. [8] In case of a 'service registry', the repository can also be used to include QoS, SLA and the current state of a service. [9]

Voice user interfaces

Voice user interfaces may have low discoverability if users are not aware of the commands that they are able to say, so these interfaces may display a list of available commands to help users find them. [10]

Metadata types

Functional

This is the basic type of meta-information that expresses the functional context of the service and the details about the product, content, or service's capabilities. The application of the standardized service contract principle helps to create the basic functional meta-data in a consistent manner. The same standardization should be applied when the same meta-information is being outside the technical contract [11] of the service e.g. when publishing information to a service registry. [12]

For general items, the data that might be used to categorize them may include:

Quality of service

For services, to know about the service behavior and its limitations, [13] and about the user experience, all of this information needs to be documented within the service registry. This way potential consumers can use this meta-information by comparing it against their performance requirements.

Considerations

Services

The effective application of this design principle requires that the meta-information recorded against each service needs to be consistent and meaningful. This is only possible if organization-wide standards exist that enforce service developers to record the required meta-data in a consistent way. The information recorded as the meta-data for the service needs to be presented in a way so that both technical and non-technical IT experts can understand the purpose and the capabilities of the service, as an evaluation of the service may be required by the business people before the service is authorized to be used.

This principle is best applied during the service-oriented analysis phase as during this time, all the details about the service's purpose and functionality are available. Although most of the service design principles support each other in a positive manner, however, in case of service abstraction and service discoverability principle, there exists an inversely proportional relationship. This is because as more and more details about the service are hidden away from the service consumers, less discoverable information is available for discovering the service. This could be addressed by carefully recording the service meta-information so that the inner workings of the service are not documented within this meta-information.

Algorithms

In the online economy, sophisticated algorithms are required for the analysis of the ways that end users search for, access and use different content or products online. Thus, not only is metadata created regarding the content or product, but also data about specific users' interaction with this content. If a social media website has a user profile for a given person, indicating demographic information (age, gender, location of residence, employment status, education, etc.), then this website can collect and analyse information about tendencies and preferences of a given user or a subcategory of users. This raises potential privacy concerns.

Algorithms have been called “black boxes”, because the factors used by the leading websites in their algorithms are typically proprietary information which is not released to the public. While a number of search engine optimization (SEO) firms offer the services of attempting to increase the ranking of a client's web content or website, these SEO firms do not typically know the exact algorithms used by Google and Facebook. Web crawlers can only access 26% of new online content "...by recrawling a constant fraction of the entire web". [14]

One concern raised with the increasing role of algorithms in search engines and databases is the creation of filter bubbles. To give a practical example, if a person searches for comedy movies online, a search engine algorithm may start mainly recommending comedies to this user, and not showing him or her the range of other films (e.g., drama, documentary, etc.). On the positive side, if this person only likes comedy films, then this restricted "filter" will reduce the information load of scanning through vast numbers of films. However, various cultural stakeholders have raised concerns about how these filter algorithms may restrict the diversity of material that is discoverable to users. Concerns about the dangers of "filter bubbles" have been raised in regards to online news services, which provide types of news, news sources, or topics to a user based on his/her previous online activities. Thus a person who has previously searched for Fox TV content will mainly be shown more Fox TV content and a person who has previously searched for PBS content will be shown more PBS search results, and so on. This could lead to news readers becoming only aware of a certain news source's viewpoints.

The search behaviour of video content viewers has changed a great deal with increasing popularity of video sharing websites and video streaming. Whereas a typical TV show consumer of the 1980s would read a print edition of TV Guide to find out what shows were on, or click from channel to channel ("channel surfing") to see if any shows appealed to them, in the 2010s, video content consumers are increasingly watching on screens (either smart TVs, tablet computer screens or smartphones), [15] that have a computerized search function and often automated algorithm-created suggestions for the viewer. With this search function, a user can enter the name of a TV show, producer, actor, screenwriter or genre to help them find content of interest to them. If the user is using a search engine on a smart device, this device may transmit information about the user's preferences and previous online searches to the website. Furthermore, in the 1980s, the type or brand of television a user was watching on did not affect his/her viewing habits. However, a person searching for TV shows in the 2010s on different brands of computerized smart TVs will probably get different search results for the same search term.

Limitations

For organizations that are trying to get maximal user uptake of their product, discoverability has become an important goal. However, achieving discovery does not automatically translate into market success. For example, if the hypothetical online game "xyz" is easily discoverable, but it will not function on most mobile devices, then this video game will not perform well in the mobile game market, despite it being at the top of search results. As well, even if the product functions, that is it runs or plays properly, as well, users may not like the product.

In the case that a user does like a certain online product or service, the discoverability has to be repeatable. If the user cannot find the product or service on a subsequent search, she or he may no longer look for this product/service, and instead shift to a substitute that is easily and reliably findable. It is not enough to make the online product or service discoverable for only a short period, unless the goal is only to create “viral content" as part of a short-term marketing campaign.

See also

Related Research Articles

Meta elements are tags used in HTML and XHTML documents to provide structured metadata about a Web page. They are part of a web page's head section. Multiple Meta elements with different attributes can be used on the same page. Meta elements can be used to specify page description, keywords and any other metadata not provided through the other head elements and attributes.

Spamdexing is the deliberate manipulation of search engine indexes. It involves a number of methods, such as link building and repeating unrelated phrases, to manipulate the relevance or prominence of resources indexed in a manner inconsistent with the purpose of the indexing system.

A web portal is a specially designed website that brings information from diverse sources, like emails, online forums and search engines, together in a uniform way. Usually, each information source gets its dedicated area on the page for displaying information ; often, the user can configure which ones to display. Variants of portals include mashups and intranet dashboards for executives and managers. The extent to which content is displayed in a "uniform way" may depend on the intended user and the intended purpose, as well as the diversity of the content. Very often design emphasis is on a certain "metaphor" for configuring and customizing the presentation of the content and the chosen implementation framework or code libraries. In addition, the role of the user in an organization may determine which content can be added to the portal or deleted from the portal configuration.

Search engine optimization (SEO) is the process of improving the quality and quantity of website traffic to a website or a web page from search engines. SEO targets unpaid traffic rather than direct traffic or paid traffic. Unpaid traffic may originate from different kinds of searches, including image search, video search, academic search, news search, and industry-specific vertical search engines.

Internet research is the practice of using Internet information, especially free information on the World Wide Web, or Internet-based resources in research.

<span class="mw-page-title-main">Metasearch engine</span> Online information retrieval tool

A metasearch engine is an online information retrieval tool that uses the data of a web search engine to produce its own results. Metasearch engines take input from a user and immediately query search engines for results. Sufficient data is gathered, ranked, and presented to the users.

Findability is the ease with which information contained on a website can be found, both from outside the website and by users already on the website. Although findability has relevance outside the World Wide Web, the term is usually used in that context. Most relevant websites do not come up in the top results because designers and engineers do not cater to the way ranking algorithms work currently. Its importance can be determined from the first law of e-commerce, which states "If the user can’t find the product, the user can’t buy the product." As of December 2014, out of 10.3 billion monthly Google searches by Internet users in the United States, an estimated 78% are made to research products and services online.

Enterprise content management (ECM) extends the concept of content management by adding a timeline for each content item and, possibly, enforcing processes for its creation, approval, and distribution. Systems using ECM generally provide a secure repository for managed items, analog or digital. They also include one methods for importing content to bring manage new items, and several presentation methods to make items available for use. Although ECM content may be protected by digital rights management (DRM), it is not required. ECM is distinguished from general content management by its cognizance of the processes and procedures of the enterprise for which it is created.

A metadata registry is a central location in an organization where metadata definitions are stored and maintained in a controlled method.

A video search engine is a web-based search engine which crawls the web for video content. Some video search engines parse externally hosted content while others allow content to be uploaded and hosted on their own servers. Some engines also allow users to search by video format type and by length of the clip. The video search results are usually accompanied by a thumbnail view of the video.

Local search is the use of specialized Internet search engines that allow users to submit geographically constrained searches against a structured database of local business listings. Typical local search queries include not only information about "what" the site visitor is searching for but also "where" information, such as a street address, city name, postal code, or geographic coordinates like latitude and longitude. Examples of local searches include "Hong Kong hotels", "Manhattan restaurants", and "Dublin car rental". Local searches exhibit explicit or implicit local intent. A search that includes a location modifier, such as "Bellevue, WA" or "14th arrondissement", is an explicit local search. A search that references a product or service that is typically consumed locally, such as "restaurant" or "nail salon", is an implicit local search.

Search Engine Results Pages (SERP) are the pages displayed by search engines in response to a query by a user. The main component of the SERP is the listing of results that are returned by the search engine in response to a keyword query.

Multimedia search enables information search using queries in multiple data types including text and other multimedia formats. Multimedia search can be implemented through multimodal search interfaces, i.e., interfaces that allow to submit search queries not only as textual requests, but also through other media. We can distinguish two methodologies in multimedia search:

Image meta search is a type of search engine specialised on finding pictures, images, animations etc. Like the text search, image search is an information retrieval system designed to help to find information on the Internet and it allows the user to look for images etc. using keywords or search phrases and to receive a set of thumbnail images, sorted by relevancy.

<span class="mw-page-title-main">Geotargeting</span> Website content based on a visitors location

In geomarketing and internet marketing, geotargeting is the method of delivering different content to visitors based on their geolocation. This includes country, region/state, city, metro code/zip code, organization, IP address, ISP, or other criteria. A common usage of geotargeting is found in online advertising, as well as internet television with sites such as iPlayer and Hulu. In these circumstances, content is often restricted to users geolocated in specific countries; this approach serves as a means of implementing digital rights management. Use of proxy servers and virtual private networks may give a false location.

Documentum is an enterprise content management platform, now owned by OpenText, as well as the name of the software company that originally developed the technology. EMC acquired Documentum for US$1.7 billion in December, 2003. The Documentum platform was part of EMC's Enterprise Content Division (ECD) business unit, one of EMC's four operating divisions.

<span class="mw-page-title-main">Metadata</span> Data about data

Metadata is "data that provides information about other data", but not the content of the data itself, such as the text of a message or the image itself. There are many distinct types of metadata, including:

<span class="mw-page-title-main">Reverse image search</span> Content-based image retrieval

Reverse image search is a content-based image retrieval (CBIR) query technique that involves providing the CBIR system with a sample image that it will then base its search upon; in terms of information retrieval, the sample image is very useful. In particular, reverse image search is characterized by a lack of search terms. This effectively removes the need for a user to guess at keywords or terms that may or may not return a correct result. Reverse image search also allows users to discover content that is related to a specific sample image or the popularity of an image, and to discover manipulated versions and derivative works.

An online video platform (OVP) enables users to upload, convert, store, and play back video content on the Internet, often via a private server structured, large-scale system that may generate revenue. Users will generally upload video content via the hosting service's website, mobile or desktop application, or other interfaces (API), and typically provides embed codes or links that allow others to view the video content.

Local search engine optimization is similar to (national) SEO in that it is also a process affecting the visibility of a website or a web page in a web search engine's unpaid results often referred to as "natural", "organic", or "earned" results. In general, the higher ranked on the search results page and more frequently a site appears in the search results list, the more visitors it will receive from the search engine's users; these visitors can then be converted into customers. Local SEO, however, differs in that it is focused on optimizing a business's online presence so that its web pages will be displayed by search engines when users enter local searches for its products or services. Ranking for local search involves a similar process to general SEO but includes some specific elements to rank a business for local search.

References

  1. Schwarzner, William W. (1988). "The Federal Rules, the Adversary Process, and Discovery Reform". University of Pittsburgh Law Review. 50: 703. Retrieved 30 September 2017.
  2. "Discoverability Summit". Discoverability Summit. Canadian Radio-television and Telecommunications Commission. Retrieved 18 February 2016.
  3. "SEO SEM: Search engine optimization/marketing" . Retrieved 14 March 2018.
  4. The overall purpose of the service
  5. "Service Contract". Archived from the original on 2012-05-01. Retrieved 2010-03-02.
  6. Michael Poulin. Evolution of principles of Service Orientation: Service Composability and Discoverability, part 7. Date accessed: 20 April 2010.
  7. Reddy, et al. Evaluating legacy assets in the context of migration to SOA. pp 58. Date accessed: 20 April 2010.
  8. Dennis Wisnosky.Principles and Patterns at the U.S. Department of Defense. Date Accessed: 20 April 2010.
  9. .Vinod Sarma, Srinivas Rao Bhagavatula. Freeway patterns for SOA systems. Date accessed: 28 April 2010.
  10. Corbett, Eric; Weber, Astrid (September 2016). What can I say?: addressing user experience challenges of a mobile voice user interface for accessibility. Proceedings of the 18th International Conference on Human-Computer Interaction with Mobile Devices and Services. pp. 72–82. doi: 10.1145/2935334.2935386 .
  11. "technical contract". Archived from the original on 2012-05-01. Retrieved 2010-03-02.
  12. A repository that contains meta-data about services in a specific format e.g. classification of service, its location, etc.
  13. "SOA Magazine - Essential Components of an SOA Quality Foundation". 5 March 2012. Archived from the original on 2012-03-05.
  14. Dasgupta, Anirban; Ghosh, Arpita; Kumar, Ravi; Olston, Christopher; Pandey, Sandeep; Tomkins, Andrew (2007). "The discoverability of the web" (PDF). Proceedings of the 16th international conference on World Wide Web. pp. 421–430. doi:10.1145/1242572.1242630. ISBN   9781595936547. S2CID   783693.
  15. Clifford, Stephanie (2008-12-08). "TV Guide, Once Master of the Airwaves, Tries to Survive in a Competitive Category". The New York Times. ISSN   0362-4331 . Retrieved 2022-12-07.