Pulse 2025 Product Roundup: From Monitoring to AI-Native Control Plane

Read more

The Future of Databases: Why the Agentic AI Era Makes Data Infrastructure More Important Than Ever

There is a common misconception that AI will eventually make databases less important — that models will simply "know" everything, or that context windows will grow large enough to hold whatever an application needs. The opposite is happening. The rise of agentic AI is creating the most demanding database requirements the industry has ever seen, and every major database category is growing as a direct result.

The global database market is tracking toward $329 billion by 2031 at nearly 14% CAGR. That growth is not driven by traditional enterprise software — it is driven by AI.

Why Agents Need Databases More Than Traditional Applications Do

A conventional web application makes a handful of database queries per user interaction: fetch a record, write a result, done. An AI agent is fundamentally different.

Agents execute multi-step tasks that unfold over time. They must remember what they have done, what they have learned, and what state they left behind. They call tools at high frequency and must store and query the structured results. They reason over relationships — between entities, documents, facts, and prior decisions. And in multi-agent architectures, they must share state with other agents running in parallel.

Every one of these requirements maps to a database problem:

  • Persistent state — context windows are ephemeral; databases are not
  • Semantic memory — past interactions must be encoded, stored, and retrieved by meaning, not just by key
  • Structured facts — tool call results, entity attributes, and world knowledge must be queryable
  • Relationship reasoning — agents must traverse connections between entities across multiple hops
  • High-frequency retrieval — agents query databases far faster than human users do, with tighter latency tolerances
  • Observability — agent behavior must be logged, traced, and analyzed at scale

The result is that a production AI agent system in 2025 typically requires not one database but five: a relational store for transactional state, a vector store for semantic retrieval, a graph database for relationship memory, an analytics database for observability, and a cache for low-latency lookups. The database layer has not been simplified — it has been multiplied.

PostgreSQL: The Anchor of OLTP in the AI Era

PostgreSQL has become the dominant database for new application development, and the AI wave has accelerated that consolidation. The 2025 Stack Overflow Developer Survey put PostgreSQL adoption at 55.6% — a 7 percentage point jump in a single year, the largest single-year gain in PostgreSQL's history. Among AI practitioners specifically, that figure rises to 59.5% according to JetBrains' 2025 data.

The primary driver is pgvector, the PostgreSQL extension that adds vector storage and approximate nearest-neighbor search directly to the database. For a large class of AI applications — RAG pipelines with moderate vector counts, semantic search on datasets up to tens of millions of records — pgvector eliminates the need for a separate dedicated vector database. One database, one operational surface, full SQL alongside semantic search.

The pgvector ecosystem has matured rapidly. Timescale's pgvectorscale benchmarks show performance competitive with purpose-built vector databases at mid-scale: 471 queries per second at 99% recall on 50 million vectors. Supabase, built entirely on PostgreSQL, reached a $5 billion valuation by late 2025 and reports that 30% of new signups are AI builders using pgvector. They now provision 2,500 new databases every day.

The practical heuristic that has emerged: use pgvector up to roughly 10 million vectors or for teams that want to minimize infrastructure complexity; use a dedicated vector database for larger workloads requiring high QPS, fine-grained multi-tenancy, or specialized indexing.

PostgreSQL's advantage is not just technical. It carries thirty years of operational knowledge, a mature ecosystem of extensions, and deep integration with every major cloud provider and application framework. AI builders who choose PostgreSQL as their OLTP foundation can incrementally add capabilities — vector search, time-series via TimescaleDB, JSON documents — without switching databases. That composability is a significant advantage in a space where requirements evolve quickly.

Vector Databases: The Infrastructure of AI Memory

Vector databases store high-dimensional numerical embeddings — the mathematical representations that machine learning models use to encode meaning. Querying a vector database is how an AI system answers the question "what is semantically similar to this?", which is the core operation behind retrieval-augmented generation (RAG), recommendation, semantic search, and multimodal similarity.

The market is growing accordingly. From roughly $1.7 billion in 2023, the vector database market is projected to reach $8.95 billion by 2030 at 27.5% CAGR. Purpose-built vector databases — Pinecone, Weaviate, Qdrant, Milvus — are differentiated by the workloads pgvector cannot yet handle comfortably: hundreds of millions to billions of vectors, extreme QPS requirements, multi-tenant isolation at scale, and fine-grained access control per collection.

Each has established a foothold: Weaviate has crossed 1 million Docker pulls per month and raised $50 million in 2024; Qdrant raised $28 million in early 2024 focused on high-performance on-premise deployments; Milvus/Zilliz maintains approximately 25,000 GitHub stars as the leading open-source option.

The more consequential shift is architectural. The era of "just add a vector database" as a standalone bolt-on is giving way to hybrid stores — databases that combine vector search with filtering, keyword search, graph traversal, or relational queries in a single system. The reason is that pure vector retrieval, in isolation, often returns the right neighborhood but the wrong answer. Production RAG systems need to filter by date, ownership, or category alongside semantic similarity. This is pushing vector capabilities into every major database category rather than remaining the exclusive domain of specialized vendors.

OLAP Databases: Built for the Speed Agents Demand

Online analytical processing databases — designed for aggregating and querying large volumes of data fast — are seeing a new category of customer: AI systems themselves.

Human analysts run queries interactively, at human speed, and tolerate latencies in the seconds-to-minutes range. AI agents run analytical queries at machine speed: continuously, automatically, and at latencies that must be in milliseconds for real-time decisions. This shifts the requirements for analytical databases significantly toward higher throughput and lower latency at scale.

ClickHouse

ClickHouse has become the analytical database of choice for AI infrastructure. In May 2025, it raised $350 million in a Series C round. Seven months later, in January 2026, it raised a further $400 million Series D at approximately a $15 billion valuation. Cloud ARR grew more than 250% year-over-year, with the customer base tripling to over 3,000 from mid-2024 to early 2026.

The customer list maps directly to the AI infrastructure ecosystem: Anthropic, Weights & Biases, LangChain, Poolside, Sierra, alongside larger enterprises including Tesla, Capital One, and Meta. ClickHouse's January 2026 Series D also included the acquisition of Langfuse, a leading open-source LLM observability platform — a direct move to become the analytics backend for teams monitoring production AI applications at scale, where high-volume trace ingestion and fast aggregations over billions of events are the core workload.

The underlying value proposition is unchanged: ClickHouse processes analytical queries an order of magnitude faster than general-purpose databases at a fraction of the infrastructure cost. In an agentic world, where queries are generated by machines rather than humans, that price-performance ratio becomes even more attractive.

DuckDB

DuckDB represents a different end of the analytical spectrum: embedded, in-process, requiring no server. Its adoption doubled in a single year on the Stack Overflow survey — from 1.4% in 2024 to 3.3% in 2025 — and it was named one of the top-three most admired databases by developers. The 1.0.0 release in mid-2024, with stable on-disk format and backward compatibility guarantees, marked its transition from interesting project to production-ready tool.

DuckDB's niche is local analytical computation: data engineering workflows, embedded analytics in applications, and lightweight pipelines where spinning up a cluster is unnecessary overhead. In AI applications, it surfaces as the engine for fast local processing of retrieved documents, structured tool outputs, or intermediate analytical steps that do not need to touch a remote database.

Snowflake and the Broader Market

The OLAP market broadly continues to grow. Snowflake reported $3.6 billion in revenue for FY2025, a 29% year-over-year increase, and has oriented its product strategy squarely around AI workloads through its Cortex LLM integration and AI Data Cloud positioning. The real-time OLAP segment specifically — databases optimized for sub-second query latency at high ingest rates — is projected to grow at 20% CAGR through 2033, driven substantially by AI-adjacent observability, event analytics, and agent telemetry workloads.

Graph Databases: The Memory Architecture for Agentic Systems

Of all the database categories accelerated by the AI shift, graph databases have arguably the most direct structural fit with what agentic AI requires.

A graph database stores data as nodes (entities) and edges (relationships), with properties on both. This maps naturally onto the way agents reason: not by looking up individual facts, but by traversing chains of relationships — who knows whom, what document references what concept, which actions led to which outcomes. Flat tables and vector indexes can retrieve relevant items; graph traversal can answer questions that require synthesizing multiple relationships across multiple hops.

The market reflects this fit. From $2.86 billion in 2024, the graph database market is projected to reach $14.58 billion by 2032 at 22.6% CAGR. Gartner projected that by 2025, graph technologies would be used in 80% of data and analytics innovations.

Neo4j, the dominant player with 44% market share in graph DBMS and presence in 84% of Fortune 100 companies, crossed $200 million in ARR in November 2024 — having doubled revenue over three years. Its stated strategic goal is to become the default knowledge layer for agentic systems, with Aura Agent enabling customers to build agents on their own graph data. Its MCP (Model Context Protocol) server integration, released shortly after Anthropic introduced the standard, connects graph databases directly into the tool-use layer of AI agents.

The academic and research framing is equally clear. Microsoft Research's GraphRAG, published and open-sourced in 2024, demonstrated that using a knowledge graph as the retrieval layer for RAG — rather than flat vector search — significantly improves the quality of answers to questions requiring synthesis across many documents. A Data.world study measured an average 3x improvement in LLM response accuracy using GraphRAG versus standard RAG across 43 business questions. The mechanism is intuitive: vector search finds documents that are semantically similar; graph traversal finds the path of relationships that explains why they are related and what connects them.

For agent memory specifically, graph databases solve the problem of structured episodic memory: storing not just what happened, but who was involved, what entities were affected, and how those entities relate to others the agent has encountered. Tools like Graphiti implement knowledge graph-based agent memory with hybrid retrieval — semantic embeddings, BM25 keyword search, and direct graph traversal — achieving P95 retrieval latency around 300ms for production agent workloads.

The Shape of the Stack

The implication of all this is that AI systems are not replacing the need for diverse database infrastructure — they are requiring more of it, more deliberately composed. A production agentic system in 2025 typically draws on:

  • PostgreSQL for user data, session state, and transactional operations
  • A vector store (pgvector for moderate scale, Qdrant or Weaviate for larger) for semantic memory and RAG retrieval
  • A graph database for knowledge representation and multi-hop relational reasoning
  • An OLAP database (ClickHouse, DuckDB, or Snowflake) for agent observability, telemetry, and analytical queries
  • A cache (Redis, Valkey) for high-frequency, low-latency lookups

This is not redundancy — each layer handles queries that the others cannot answer efficiently. The engineering discipline of choosing the right database for each workload, rather than forcing a single system to do everything, has become more important in the AI era, not less.

The industry shorthand for this is polyglot persistence, and it has been the right architectural instinct for a decade. What has changed is the forcing function: AI applications expose the limitations of any single database more quickly than traditional applications did, because agents query more frequently, across more data modalities, and with less tolerance for latency.

What This Means

The database market is entering its most consequential growth phase. Categories that were niche five years ago — vector search, graph traversal, embedded OLAP — are now infrastructure primitives that AI teams evaluate from day one. Categories that were already dominant — PostgreSQL, cloud OLAP — are growing faster than they were before AI.

The teams building AI agents are learning, often the hard way, that the intelligence layer is only as useful as the data layer beneath it. An agent without durable memory forgets. An agent without fast retrieval is slow. An agent without relational context hallucinates connections that do not exist. The database is not an implementation detail of an AI system — it is the foundation that determines what the system can know, remember, and reason about.

The data infrastructure decisions made now will determine the capabilities — and the limitations — of the AI systems built on top of them.

  • Vector Databases
  • Graph Databases
  • PostgreSQL
  • ClickHouse
  • Retrieval-Augmented Generation (RAG)
  • AI Agents
  • Knowledge Graphs
  • OLAP Databases
Pulse - Elasticsearch Operations Done Right

Pulse can solve your Elasticsearch issues

Subscribe to the Pulse Newsletter

Get early access to new Pulse features, insightful blogs & exclusive events , webinars, and workshops.

We use cookies to provide an optimized user experience and understand our traffic. To learn more, read our use of cookies; otherwise, please choose 'Accept Cookies' to continue using our website.