Back to climacs.net
Interactive Infographic

Vector Databases Explained

The infrastructure behind modern AI search

๐Ÿ—„๏ธ
Traditional DB
Rows & Columns
โ†’
๐Ÿ”ฎ
Vector DB
Semantic Space
๐ŸŽฏ
Better Retrieval
Search by meaning, not keywords
โšก
Production Scale
Billions of vectors, milliseconds
๐Ÿค–
Grounded AI
LLM answers backed by your data
Panel 1

Why Keywords Fall Short

Traditional search matches words. AI needs to match meaning.

๐Ÿ” User searches:

"How to fix a leaky pipe?"

โŒ Keyword Search

Misses "Resolving Residential Plumbing Failures" โ€” the tokens don't overlap.

Words must match exactly

โœ… Vector Search

Finds it โ€” embeddings encode that "leaky pipe" and "plumbing failure" mean the same thing.

Meaning is what matters

This is the keyword gap โ€” and it's fundamental to why vector databases exist.

Panel 2

Same Word. Five Meanings.
Vector Search Knows.

The word "bank" appears in every sentence โ€” but only one answers the question.

๐Ÿ” Query:

"Where can I store my money?"

โŒ Geography
"The bank of the river is prone to erosion during the spring thaw."
โœ… Match โ€” Finance
"I need to visit the local bank to deposit my paycheck before it closes."
โŒ Aviation
"The pilot had to bank the aircraft sharply to avoid the storm clouds."
โŒ Idiom
"You can always bank on her to be the first person in the office."
โŒ Technology
"The technician replaced a faulty module in the server bank."

Keyword search can't distinguish these. Vector search understands context.

Panel 3

What Is a Vector?

Think GPS, but for meaning instead of location.

๐ŸŒ GPS Analogy

GPS: [48.86, 2.35] โ†’ Paris
Embedding: [0.82, 0.15, 0.91, ...]
โ†’ "The cat sat on the mat"

A GPS coordinate is 2 numbers describing physical location. An embedding is 1,536 numbers describing location in meaning-space.

๐Ÿ“ Semantic Space

cat
kitten
dog
car
truck
Simplified 2D projection

Words with similar meaning cluster together โ€” cat, kitten, and dog are close because they're all animals. Car and truck form their own cluster โ€” similar to each other, but far from the animals.

The distance between dots = how different their meanings are. This is how vector search finds relevant results without matching keywords.

โ–ผ WANT TO GO DEEPER? โ–ผ

The engineering deep-dive starts here

Panel 4 โ€” Engineering

How Vector DBs Make It Fast

Approximate Nearest Neighbor (ANN) trades tiny accuracy loss for massive speed gains.

๐Ÿ“š The Library Analogy

Imagine a library with 1 million books. You want the 5 most similar to yours.

Brute Force

Read every book and compare. Correct, but takes years.

ANN (What Vector DBs Do)

Organize into neighborhoods. Walk to the right section, compare nearby books only. Done in seconds.

๐Ÿ•ธ๏ธ

HNSW โ€” Graph Navigation

Vectors connected in a multi-layer graph. Search hops between neighbors, getting closer each step. Fast, accurate, most widely used.

๐ŸŽฏ

IVF โ€” Cluster Search

Vectors grouped into clusters. Query compares to cluster centers first, then searches only the closest clusters. Tunable speed/accuracy.

โšก Latency at Scale

5ms
6ms
7ms
8ms
HNSW โ€” Stays flat
10ms
100ms
5s
min+
Brute Force โ€” Explodes
1K vectors100K10M1B
Panel 5 โ€” The Core Diagram

RAG Pipeline Architecture

Retrieval-Augmented Generation: how AI answers questions from your documents.

๐Ÿ“„ Ingestion Phase

๐Ÿ“„ Documents
โ†’
โœ‚๏ธ Chunking
โ†’
๐Ÿงฎ Embed
โ†’
๐Ÿ—„๏ธ Vector DB

๐Ÿ” Query Phase

โ“ User Question
โ†’
๐Ÿงฎ Query Embed
โ†’
๐Ÿ”Ž Topโ€‘k Retrieval
โ†’
๐Ÿค– LLM
โ†’
๐Ÿ’ฌ Answer

โš ๏ธ Bad Chunking

Too small = missing context. Too large = diluted relevance. Test multiple strategies.

โš ๏ธ Weak Embeddings

A general model on domain-specific text retrieves poorly. Use domain-tuned models.

โš ๏ธ No Reranking

Top-k by similarity โ‰  best for answering. Add a reranker for production quality.

Panel 6

Choosing a Vector Database

A decision tree, not a comparison table โ€” pick the path that fits your situation.

Already use PostgreSQL?โ†’ pgvector
Need zero ops at massive scale?โ†’ Pinecone
Open source + hybrid search?โ†’ Weaviate
Max performance + billion-scale, open source?โ†’ Milvus / Qdrant
Just prototyping locally?โ†’ Chroma
Multimodal (images + text + audio)?โ†’ LanceDB / Weaviate
๐Ÿ’ก Pro tip: For retrieval quality, your embedding model matters more than your database choice. For retrieval speed and cost, the database decision is critical.
Panel 7

What Actually Matters

Three things people overlook โ€” and they're more important than which database you pick.

1. Embedding Model > Database Choice

A strong embedding model with a simple index will outperform a weak model on the fanciest database. Invest in embedding quality first.

2. Chunking Strategy Is a Design Decision

There's no universal chunk size. It depends on your data, queries, and context window. Test 256, 512, 1024 tokens with overlap.

3. Hybrid Retrieval Is Table Stakes

Production systems combine vector similarity + keyword/BM25 + metadata filters โ†’ final ranking. Don't rely on vector search alone.

๐Ÿ”€ Hybrid Search in Practice

Vector Search
+
Keyword/BM25
+
Metadata Filters
โ†’
Combined Ranking

๐Ÿ“‹ Getting Started Checklist

  • Select and test an embedding model on your data
  • Determine chunk size with overlap (512? 1024?)
  • Design your metadata schema (time, type, permissions)
  • Configure hybrid search (BM25 + vector weights)
  • Evaluate a reranker (does it improve your top-k?)
  • Test recall and latency at expected scale

๐Ÿ—๏ธ OUR STACK โ€” "Everything above, we run it."

Vector DBMilvus (HNSW, open source)
Embeddingsbge-small-en-v1.5 (384-dim)
RerankerGTE-multilingual
LLM ProxyLiteLLM โ†’ Bedrock / Azure OpenAI
Search TypeHybrid (vector + metadata)
Embedding โ€” converting text/image to a list of numbers
ANN โ€” Approximate Nearest Neighbor (fast, ~99% accurate)
HNSW โ€” Hierarchical Navigable Small World (graph-based ANN)
IVF โ€” Inverted File Index (cluster-based ANN)
Cosine Similarity โ€” angle between vectors (direction matters)
RAG โ€” Retrieval-Augmented Generation
Top-k โ€” the k best search results (e.g., top-5)
BM25 โ€” traditional keyword relevance scoring

Climacs IT Consulting ยท Based on "Vector Databases Explained" by Darren Broemmer ยท March 2026

ยฉ 2026 Climacs IT Consulting