Vector Search
Vector search (or similarity search) finds semantically similar items by comparing high-dimensional vector embeddings. It is foundational for semantic search, recommendation systems, and retrieval-augmented-generation-rag.
Core Mechanisms
- Vector Embeddings: Numerical representations of data (text, images) in a continuous vector space (e.g., via BERT, Sentence Transformers).
- Similarity Metrics: Cosine similarity or Euclidean distance to measure vector proximity.
- Approximate Nearest Neighbor (ANN) Algorithms: Efficiently search large vector databases (e.g., FAISS, HNSW).
Limitations and Alternatives
- Embedding Consistency Requirement: Traditional vector search requires the same model for both embedding generation and retrieval (e.g., using
text-embedding-ada-002for both steps). - Graph RAG: A flexible alternative using knowledge graphs and LLMs for structured retrieval. Does not require consistent models for embedding vs. retrieval, enabling more adaptable systems.
- IBM Explainer: Creating GraphRAG (video demonstration)
- Key advantage: Graph structures allow explicit relationship traversal (e.g., “find all products related to
batteryviamanufacturer”).
Applications
- Semantic search in enterprise knowledge bases
- Personalized recommendations
- Contextual LLM augmentation via rag
2026 04 14 IBM Explainer creating GraphRAG
Source Notes
- 2026-04-14: [[lab-notes/2026-04-14-Optimizing-AI-Costs-and-Privacy-with-Local-Open-Source-Models-and-Hybr|“But OpenClaw is expensive…“]]