Semantic search denotes search with meaning, as distinguished from lexical search where the search engine looks for literal matches of the query words or variants of them, without understanding the overall meaning of the query. [1] Semantic search seeks to improve search accuracy by understanding the searcher's intent and the contextual meaning of terms as they appear in the searchable dataspace, whether on the Web or within a closed system, to generate more relevant results.
Some authors regard semantic search as a set of techniques for retrieving knowledge from richly structured data sources like ontologies and XML as found on the Semantic Web. [2] Such technologies enable the formal articulation of domain knowledge at a high level of expressiveness and could enable the user to specify their intent in more detail at query time. [3] The articulation enhances content relevance and depth by including specific places, people, or concepts relevant to the query. [4]
Tools like Google's Knowledge Graph provide structured relationships between entities to enrich query interpretation. [5]
Models like BERT and Sentence-BERT convert words or sentences into dense vectors for similarity comparison. [6]
Semantic ontologies like Web Ontology Language, Resource Description Framework, and Schema.org organize concepts and relationships, allowing systems to infer related terms and deeper meanings. [7]
Hybrid search models combine lexical retrieval (e.g., BM25) with semantic ranking using pretrained transformer models for optimal performance. [8]
Term Frequency–Inverse Document Frequency (TF–IDF) is a statistical measure used to evaluate the importance of a word in a document relative to a collection (corpus). It is widely used for information retrieval and text mining. [9]
Latent Semantic Analysis (LSA) is a technique in natural language processing that uncovers hidden relationships between words by reducing the dimensionality of the term-document matrix using Singular Value Decomposition (SVD). [10]
Word2Vec is a neural network-based model that learns continuous vector representations of words (embeddings) by predicting surrounding words (context). [11]
GloVe is an unsupervised learning algorithm for generating word embeddings by leveraging word co-occurrence statistics across a large corpus. [12]
BERT is a transformer-based model that pre-trains deep bidirectional representations by conditioning on both left and right context in all layers. [13]
Sentence-BERT (S-BERT) extends BERT by fine-tuning it with a Siamese and triplet network structure to generate semantically meaningful sentence embeddings. [14]
Cosine Similarity measures the cosine of the angle between two vectors, commonly used for text similarity and recommendation systems. [15]
Euclidean Distance is the straight-line distance between two vectors in a multidimensional space. [16]
Dot-Product Similarity calculates similarity between vectors as the inner product, often used in neural networks and recommender systems. [17]
Approximate Nearest Neighbor (ANN) search algorithms aim to retrieve points close to a query point in high-dimensional space while trading off exact accuracy for speed and scalability. [18]
Hierarchical Navigable Small World (HNSW) is a graph-based ANN method that constructs a hierarchical, navigable small-world graph to achieve logarithmic search complexity. [19]
Locality-Sensitive Hashing (LSH) is a method that hashes input vectors into buckets so that similar items are more likely to map to the same bucket, enabling sub-linear search time. [20]
Inverted File Index (IVF) is a traditional indexing technique where a mapping is maintained from content (words, features) to their document or vector identifiers. It is often combined with clustering (e.g., k-means) for ANN search. [21]
![]() |
![]() |
![]() |