Vector Databases

Specialized databases optimized for storing, indexing, and retrieving high-dimensional vector embeddings. Enable efficient similarity search (e.g., nearest neighbor queries) for applications like semantic search, recommendation systems, and LLM-powered retrieval.

Core Functionality

  • Approximate Nearest Neighbor (ANN) Search: Uses algorithms like HNSW, IVF, or FAISS for scalable similarity matching.
  • Embedding Support: Stores vectors generated from text, images, or other modalities via models like BERT, CLIP, or Sentence Transformers.
  • Scalability: Handles millions/billions of vectors with low-latency queries.

Key Limitations

  • Model Constraint: Requires the same embedding model for both vector generation and retrieval (e.g., BERT embeddings must be searched with BERT).
  • Semantic Rigidity: Struggles with complex relationships beyond vector similarity (e.g., hierarchical or causal links).

GraphRAG: Flexible Alternative

  • GraphRAG leverages knowledge graphs and LLMs to query structured data, eliminating the model constraint of vector search.
  • Flexibility: Uses different models for graph construction (e.g., GNNs) and retrieval (e.g., LLMs), enabling richer context.
  • Advantage: Better handles complex queries involving relationships (e.g., “Show me products similar to X that are also used with Y”) compared to pure vector similarity.

Integration Ecosystem

  • RAG Pipelines: Primary component in Retrieval-Augmented Generation for LLM context retrieval.
  • Knowledge Graphs: Knowledge Graphs enhance vector search with structured relationships (e.g., Neo4j + vector DBs).
  • Embedding Models: embedding-models must align with vector DB capabilities (e.g., dimensionality, metric).

References

  • 2026 04 14 IBM Explainer creating GraphRAG

Source Notes

  • 2026-04-23: [[lab-notes/2026-04-23-Claude-Routines-Action-Based-AI-Automation-for-Business-Event-Response|Claude Routines: Action-Based AI Automation for Business Event Response]]