ProvenData & RetrievalNo changeMarch 2026 Backfill

Battle-tested in production. Build on it with confidence.

pgvector

The pragmatic choice for teams already on Postgres who need vector search under ~10M vectors — beyond that scale, evaluate dedicated vector databases.

RAG·Infrastructure·Open-source

github.com

Our Take

What It Is

pgvector adds vector similarity search (cosine, L2, inner product) to PostgreSQL. You store embeddings in the same database as your application data, query them with SQL, and use HNSW or IVFFlat indexes for fast approximate nearest-neighbour search. Version 0.8.2 (March 2026) fixes a buffer overflow in parallel HNSW builds (CVE-2026-3172).

Why It Matters

pgvector is Proven because it solved the right problem: most teams already run Postgres, and adding vector search without another database in the stack eliminates operational complexity. The pgvectorscale benchmarks (Timescale) showed 471 QPS versus Qdrant's 41 QPS at 99% recall on 50M vectors — competitive performance without a separate system to manage.

The practical appeal is obvious. One connection string. One backup strategy. One set of access controls. For teams building RAG applications with millions (not billions) of vectors, pgvector is the path of least resistance.

Key Developments

  • Mar 2026: pgvector 0.8.2 fixes buffer overflow with parallel HNSW index builds (CVE-2026-3172).
  • Late 2025: pgvector 0.8.0 adds iterative index scans to prevent overfiltering and improve query performance with filters.
  • Ongoing: Supported on all major managed Postgres services: AWS RDS, Google Cloud SQL, Azure, Supabase, Neon, Timescale.

What to Watch

The scalability ceiling is where pgvector's story ends and dedicated vector databases begin. If your dataset grows past tens of millions of vectors or you need horizontal scaling for vector search, you'll need to evaluate alternatives. Watch for pgvector's filtered search improvements — combining vector similarity with relational filters is still rougher than what dedicated databases like Weaviate provide out of the box.

Strengths

  • Zero infrastructure overhead: Runs inside existing PostgreSQL. One connection string for relational and vector queries.
  • HNSW performance: 30x QPS boost and 30x p99 latency improvement over IVFFlat at 99% recall. Competitive under 10M vectors.
  • Broad managed service support: Available on AWS RDS, Google Cloud SQL, Azure, Supabase, Neon, and Timescale.
  • Rich type support: halfvec, sparsevec, and standard vector types cover most embedding models. 19K+ GitHub stars.

Considerations

  • Scalability ceiling: Works well for millions of vectors, but billion-scale requires sharding or external systems. No horizontal scaling.
  • HNSW index build cost: HNSW builds 12-42x slower than IVFFlat and produces larger indexes. Frequent updates are operationally expensive.
  • Filtered search complexity: Combining vector similarity with relational filters requires manual optimisation.
  • No native hybrid search: Combining full-text search with vector search requires manual implementation (reciprocal rank fusion).