Your AI doesn’t need another database

Every hot, new workload gets its own database. Briefly. You know the drill. From search to JSON (documents) to graph, as an industry we have this weird fixation with building new databases. DB-Engines now tracks 434 of them. We’re now doing it again with vector databases, which were hailed almost overnight as the essential new persistence layer for AI. The story was simple and, for a time, convincing: Traditional databases don’t understand vectors! AI applications need vectors! So AI applications must need vector databases. Right?

Nope. After all, that first premise became untrue almost immediately.

I would say that, wouldn’t I? I mean, I work for Oracle, and Oracle AI Database 26ai can store vector embeddings alongside business data, and it supports HNSW and IVF vector indexes. But it’s not just Oracle. Literally every database that developers have used for years has vector support now. Microsoft has added a native VECTOR data type to SQL Server 2025, along with vector search and vector indexes. MongoDB has pushed automated embeddings into Atlas Vector Search, with embeddings generated in the database and synchronized as data changes. Postgres, through pgvector, also offers vector support. Etc., etc., etc.

That doesn’t mean Pinecone, Weaviate, Milvus, or the other purpose-built vector vendors are doomed, but it does call into question the premise behind their VC pitch decks. For most enterprise applications, vector support is a feature, one that should be tightly woven into an existing data estate.

This matters because the hardest part of production AI isn’t nearest-neighbor search: It’s context.

Proliferating data siloes

I’m not suggesting that vector search isn’t a thing: It’s very important. If you’re building retrieval-augmented generation (RAG), recommendation systems, personalization, agent memory, or anything that requires matching meaning rather than keywords, you need some way to compare vectors efficiently. And, credit where it’s due, the purpose-built vector vendors made that obvious before the incumbents did. I was working at MongoDB when Pinecone, Weaviate, Milvus, Qdrant, and others helped establish the patterns that everyone now treats as obvious. That’s real innovation.

The problem is that it’s an incomplete innovation.

It didn’t take long for incumbents to catch on and catch up. Pinecone launched in 2021, and Postgres released pgvector that same year. Oracle, Microsoft, and others matched not long after MongoDB. Now everyone offers vector support, and the competition has moved to how well vectors are integrated into the larger product. This is particularly important in AI, where delivering context (agentic memory) depends on eliminating siloed data. It’s simply not going to work to have one system holding relational data, another with JSON documents, and still another with vectors (not to mention graph, time-series, data). Application code has to synchronize this mess while also enforcing permissions, freshness, lineage, and business rules.

That can work for a demo. It’s a nightmare in production.

Think about it. Imagine a customer updates a support article, or a compliance policy is edited. The transactional system has the new truth (hurray!), but the vector representation may not (boo!). Still, AI being AI, the model will retrieve yesterday’s context with today’s confidence. It’s fast, it’s confident, and it’s wrong.

Enterprise AI memory isn’t merely vector embeddings. It’s also documents, transactions, customer records, policies, conversations, etc. It’s also knowing what changed, who is allowed to see it, and whether the answer can be traced back to a source the business trusts. A vector database can help with part of this, but it definitely doesn’t automatically solve all of it. If vectors live in one system, documents in another, transactions in another, graph relationships somewhere else, and permissions in application code, the “context layer” becomes a distributed consistency problem with a chatbot on the front end. That’s bad.

The closer vector search gets to the governed data, the less glue code developers have to write and the fewer stale copies they have to defend. Hence, this isn’t really about “incumbents learning new tricks.” It’s about the operational center of gravity moving back to the systems that already hold enterprise truth.

When to use a standalone vector database

None of this means developers should never use a standalone vector database. That would be the same lazy thinking in the opposite direction.

Use the database you already have when vectors describe data already living there, when freshness matters, when permissions are tied to operational records, or when semantic search is one feature inside a broader application. If your AI assistant is retrieving from customer records, support tickets, product catalogs, policy documents, invoices, or account data already governed by an existing database, start there. Don’t add a second source of truth just because the tutorial did.

By contrast, reach for a standalone vector database when retrieval is the product, not simply a feature. If you’re building a search company, a recommendation platform, a large multitenant RAG service, or a system where latency, ranking sophistication, indexing behavior, or operational isolation truly exceed what your primary database can handle, then specialized infrastructure may be the right answer. But that’s not going to be the norm.

The claim “sometimes you need specialized retrieval infrastructure” is obviously true. The claim “every serious AI application needs a standalone vector database” is obviously false. The standalone vector vendors increasingly acknowledge that vector storage isn’t enough. Hence they’re moving up the stack toward retrieval quality, agent memory, latency, evaluation, data freshness, and operational simplicity.

Pinecone’s Nexus announcement is a good example. The pitch is no longer simply “store vectors here.” It’s “make AI knowledgeable.” But Pinecone’s evolution also proves the point. If the long-term moat is not storing vectors but rather delivering superior retrieval infrastructure for specialized workloads, then the standalone vector database market is smaller and more demanding than many assumed. It has to win on quality, scale, latency, developer experience, and operational simplicity. It can’t merely win by being the place where embeddings live, because embeddings increasingly live everywhere.

That’s a much harder business. Sure, it may still be a good one, but it’s no longer the default architecture for every AI team with a pile of documents and a chatbot demo. And it shouldn’t be for enterprises that need to weave vectors into their broader data estate.

Don’t start with another database

We’ve spent the past decade being told that “purpose-built databases” were the answer, when they were really just a way for one cloud vendor to ship its org chart. That advice produced architectures where simple applications somehow require five persistence layers and a team of engineers to keep them synchronized. It’s madness.

AI makes that worse because AI is hungry for context. Retrieval doesn’t happen in a vacuum. No, it happens against data with ownership, recency, permissions, lineage, and business meaning. The more systems involved, the harder it is to know whether the model is answering from truth, from an outdated copy of truth, or from something that used to be true last week.

All this is to say it’s not really about vectors. It’s about whether we want AI applications to repeat the worst habits of modern data infrastructure: more services, more pipelines, more copies, more glue code, and more places for truth to drift. Instead of creating another stale context window, we need to put vector search as close as possible to the data it describes, which usually means starting with the database you already operate and using the security model you already trust.

Sure, there’s still room for standalone vector databases, just as there’s still room for search engines, caches, and analytics platforms. Pinecone and others have a credible future if they keep proving that specialized retrieval infrastructure delivers enough value to justify another system. But the burden of proof has shifted. Developers shouldn’t ask, “Why wouldn’t I use a vector database?” and instead should ask, “Why should I add another database when the one I already have can do this well enough with significantly better performance and security across my data?”

That may be uncomfortable for a market with a lot of venture money behind it. Good. Discomfort is what happens when a feature stops looking like a category, and VCs aren’t the ones building the apps developers’ jobs depend on.

Go to Source

Author: