Negative search

Last updated

Negative search is the elimination of information which is not relevant from a mass of content in order to present to a user a range of relevant content.

Negative search is different from both positive search and discovery search. Positive search uses the selection of relevant content as its primary mechanism. Discovery calculates relatedness (between user intent and content) to present users with relevant alternatives of which they may not have been aware.

Negative search applies to those forms of searches where the user has the intention of finding a specific, actionable piece of information but lacks the knowledge of what that specific information is or might be.

Negative search can also apply to searches where the user has a clear understanding of negative intent (what they don't want) rather than what they do.

Examples of negative intent are:

- Job searching: someone knows they want a new job but they have no idea what it might be. They just know what they don't want.

- Online dating: someone is looking for a dating partner, but cannot identify what criteria they are looking for. They just know what they don't want.

- An investigator is looking for a car but has no other information on that car on which to base a search.

Negative search classifiers

If there are two forms of search (positive and negative) it follows that there are two forms of classifier models: inclusive classifiers and exclusive classifiers.

Countries of the world are a good example of a MECE list. A positive search for the country Kenya would identify content referencing Kenya and present it. A negative search for the country Kenya would exclude all content relating to other countries in the world leaving the user with content of some relevance to Kenya.

Irrelevancy as a desirable construct

Positive search tends to view irrelevancy as undesirable. Having a system actively identify and pursue irrelevant content for the purpose of elimination from a user experience may prove a highly powerful mechanism.

It follows that positive and negative search are not mutually exclusive and that a more powerful search may result from the combination of selection and elimination as tools to empower user experience in negative searches.

Degrees of passivity

Positive search involves an active search by a user with no degree of passivity (or openness). For example: "I am only interested in the Hilton Hotel in Vientiane on New Year's Eve."

Discovery involves a simultaneous secondary more Passive search by the user while they are involved in a positive search. For example: "I am interested in the Hilton Hotel in Vientiane on New Year's Eve but if there's a better hotel, let me know."

Negative search also involves an active search but with a much higher degree of passivity (or openness to discovery). For example: "I need a holiday and really don't care where as long as its good."

Searchers can be active in one dimension (positive search) while simultaneously being passive to alternatives or what they don't know they're looking for in many dimensions. In discovery they are passive in a small number of dimensions but in negative search they are passive in many or all dimensions.

Related Research Articles

A search engine is an information retrieval system designed to help find information stored on a computer system. It is an information retrieval software program that discovers, crawls, transforms, and stores information for retrieval and presentation in response to user queries. The search results are usually presented in a list and are commonly called hits. A search engine normally consists of four components, as follows: a search interface, a crawler, an indexer, and a database. The crawler traverses a document collection, deconstructs document text, and assigns surrogates for storage in the search engine index. Online search engines store images, link data and metadata for the document as well.

Search engine optimization (SEO) is the process of improving the quality and quantity of website traffic to a website or a web page from search engines. SEO targets unpaid traffic rather than direct traffic or paid traffic. Unpaid traffic may originate from different kinds of searches, including image search, video search, academic search, news search, and industry-specific vertical search engines.

Internet research is the practice of using Internet information, especially free information on the World Wide Web, or Internet-based resources in research.

<span class="mw-page-title-main">Banner blindness</span> Tendency to ignore banner-size notices

Banner blindness is a phenomenon in web usability where visitors to a website consciously or unconsciously ignore banner-like information. A broader term covering all forms of advertising is ad blindness, and the mass of banners that people ignore is called banner noise.

In text retrieval, full-text search refers to techniques for searching a single computer-stored document or a collection in a full-text database. Full-text search is distinguished from searches based on metadata or on parts of the original texts represented in databases.

A sequence profiling tool in bioinformatics is a type of software that presents information related to a genetic sequence, gene name, or keyword input. Such tools generally take a query such as a DNA, RNA, or protein sequence or ‘keyword’ and search one or more databases for information related to that sequence. Summaries and aggregate results are provided in standardized format describing the information that would otherwise have required visits to many smaller sites or direct literature searches to compile. Many sequence profiling tools are software portals or gateways that simplify the process of finding information about a query in the large and growing number of bioinformatics databases. The access to these kinds of tools is either web based or locally downloadable executables.

Hyperlocal is information oriented around a well-defined community with its primary focus directed toward the concerns of the population in that community. The term can be used as a noun in isolation or as a modifier of some other term. When used in isolation it refers to the emergent ecology of data, aggregators, publication mechanism and user interactions and behaviors which centre on a resident of a location and the business of being a resident. More recently, the term hyperlocal has become synonymous with the combined use of applications on mobile devices and GPS technology. Use of the term originated in 1991, in reference to local television news content.

Local search is the use of specialized Internet search engines that allow users to submit geographically constrained searches against a structured database of local business listings. Typical local search queries include not only information about "what" the site visitor is searching for but also "where" information, such as a street address, city name, postal code, or geographic coordinates like latitude and longitude. Examples of local searches include "Hong Kong hotels", "Manhattan restaurants", and "Dublin car rental". Local searches exhibit explicit or implicit local intent. A search that includes a location modifier, such as "Bellevue, WA" or "14th arrondissement", is an explicit local search. A search that references a product or service that is typically consumed locally, such as "restaurant" or "nail salon", is an implicit local search.

In the fields of information technology and systems management, application performance management (APM) is the monitoring and management of the performance and availability of software applications. APM strives to detect and diagnose complex application performance problems to maintain an expected level of service. APM is "the translation of IT metrics into business meaning ."

Sentiment analysis is the use of natural language processing, text analysis, computational linguistics, and biometrics to systematically identify, extract, quantify, and study affective states and subjective information. Sentiment analysis is widely applied to voice of the customer materials such as reviews and survey responses, online and social media, and healthcare materials for applications that range from marketing to customer service to clinical medicine. With the rise of deep language models, such as RoBERTa, also more difficult data domains can be analyzed, e.g., news texts where authors typically express their opinion/sentiment less explicitly.

Clean URLs are web addresses or Uniform Resource Locator (URLs) intended to improve the usability and accessibility of a website, web application, or web service by being immediately and intuitively meaningful to non-expert users. Such URL schemes tend to reflect the conceptual structure of a collection of information and decouple the user interface from a server's internal representation of information. Other reasons for using clean URLs include search engine optimization (SEO), conforming to the representational state transfer (REST) style of software architecture, and ensuring that individual web resources remain consistently at the same URL. This makes the World Wide Web a more stable and useful system, and allows more durable and reliable bookmarking of web resources.

Keyword research and optimization

Participatory culture, an opposing concept to consumer culture, is a culture in which private individuals do not act as consumers only, but also as contributors or producers (prosumers). The term is most often applied to the production or creation of some type of published media.

<span class="mw-page-title-main">Precision and recall</span> Pattern recognition performance metrics

In pattern recognition, information retrieval, object detection and classification, precision and recall are performance metrics that apply to data retrieved from a collection, corpus or sample space.

Visual networking refers to an emerging class of user applications that combine digital video and social networking capabilities. It is based upon the premise that visual literacy, "the ability to interpret, negotiate and make meaning from information presented in the form of a moving image", is a powerful force in how humans communicate, entertain and learn. The duality of visual networking—subsuming entertainment and communications, professional and personal content, video and other digital media, data networks and social networks to create immersive experiences, when, where and how the user wants it. These applications have changed video content from long-form movies and broadcast television programming to a database of segments or "clips", and social network annotations. And the generation and distribution of content takes on a new dimension with Web 2.0 applications—participatory social-networks or communities that facilitate interactive creativity, collaboration and sharing between users.

DeepPeep was a search engine that aimed to crawl and index every database on the public Web. Unlike traditional search engines, which crawl existing webpages and their hyperlinks, DeepPeep aimed to allow access to the so-called Deep web, World Wide Web content only available via for instance typed queries into databases. The project started at the University of Utah and was overseen by Juliana Freire, an associate professor at the university's School of Computing WebDB group. The goal was to make 90% of all WWW content accessible, according to Freire. The project ran a beta search engine and was sponsored by the University of Utah and a $243,000 grant from the National Science Foundation. It generated worldwide interest.

Search-based applications are software applications in which a search engine platform is used as the core infrastructure for information access and reporting. Search-based applications use semantic technologies to aggregate, normalize and classify unstructured, semi-structured and/or structured content across multiple repositories, and employ natural language technologies for accessing the aggregated information.

Personalized search refers to web search experiences that are tailored specifically to an individual's interests by incorporating information about the individual beyond the specific query provided. There are two general approaches to personalizing search results, involving modifying the user's query and re-ranking search results.

In pay-per-click advertising (PPC), negative keywords prevent advertisements from displaying for particular keyword phrases. The process of adding in negative keywords to a paid search account is an important part of the optimization process, as it allows for eliminating search queries which are unlikely to convert into a sale. For example, an individual may want to advertise for "hotel in Toronto," but does not want to display ads for the keyword "hotel jobs in Toronto." In this instance, the term "jobs" would be added as a negative keyword. There are few automated tools to predict the negative keywords and make the process simpler. Taking help from tools helps to save time and avoid the possibility of human error.

User intent, otherwise known as query intent or search intent, is the identification and categorization of what a user online intended or wanted to find when they typed their search terms into an online web search engine for the purpose of search engine optimisation or conversion rate optimisation. Examples of user intent are fact-checking, comparison shopping or navigating to other websites.

References