Join us
@kaptain ・ Jan 12,2026
Running databases on Kubernetes has moved from experimentation to standard practice, driven by platform maturity, cost pressures, and AI/ML demands. According to the 2025 Data on Kubernetes survey, organizations are now focused on operational excellence, with cost optimization, storage performance, and AI workloads shaping the next phase.
Running database workloads on Kubernetes has become a standard practice due to platform maturity, cost control, and architectural consolidation.
Organizations are increasingly adopting a hybrid model that integrates cloud and on-premises environments using Kubernetes as a universal control plane.
Cost optimization has emerged as the top priority for organizations using Kubernetes for data workloads.
The growth of AI/ML workloads is influencing the use of Kubernetes for data management.
Despite the maturity of Kubernetes for data workloads, performance bottlenecks such as storage I/O and model loading times remain challenges.
For years, the conventional wisdom was simple: run applications on Kubernetes, but leave databases to managed cloud services. That assumption is now cracking.
Industry reports and technical analyses show a clear trend: more organizations are running database workloads directly on Kubernetes, driven by platform maturity, cost control, and architectural consolidation.
Kubernetes has evolved beyond stateless orchestration. Improvements in StatefulSets, persistent volumes, operators, and cloud-native storage have made it increasingly viable to run production databases inside clusters. According to Solutions Review, databases have become one of the most common production workloads on Kubernetes, especially in data-intensive and AI-driven environments.
This shift is also reflected in broader platform engineering discussions. The New Stack notes that Kubernetes is finally addressing its long-standing weakness around persistent state, making database operations more predictable and automatable using operators and declarative workflows.
Cost pressure is another major driver. Organizations are reassessing the long-term economics of managed cloud databases, which often carry opaque pricing and scaling penalties. CIO.com reports that cloud repatriation is back on the agenda, particularly for data-heavy workloads where cost predictability and control matter more than convenience.
Rather than abandoning the cloud, many teams are converging on a hybrid model: Kubernetes as the universal control plane, hosting both applications and databases, across cloud and on-prem environments. This allows tighter integration, reduced vendor lock-in, and a single operational model for stateful and stateless systems alike.
The trend is reinforced by data from the 2025 Data on Kubernetes Community (DoKC) survey, which reports that running data workloads on Kubernetes is now a standard practice rather than an emerging one. The survey shows that organizations have largely moved past initial adoption and are now focused on operational optimization, particularly around cost management, performance, and AI/ML readiness.
Cost optimization emerged as the top priority for 2025, overtaking security and scaling initiatives. This focus is especially pronounced among organizations running AI and machine learning workloads, where storage costs have become a primary concern due to the size of training datasets, model checkpoints, and inference artifacts. The survey also highlights the growing role of vector databases, which a large majority of respondents now consider critical infrastructure for AI workloads.
The data further points to a shift in architecture. A majority of respondents identified edge computing and real-time data processing as essential to their future data strategy, reflecting growing demand for low-latency processing and compliance with data sovereignty requirements. These needs are pushing organizations toward more distributed Kubernetes-based architectures rather than centralized, batch-oriented systems.
At the same time, the survey identifies persistent challenges. Performance optimization, security and compliance, and skills shortages remain the most commonly cited obstacles. The skills gap is particularly notable, with organizations reporting difficulty finding practitioners who combine Kubernetes operational expertise with deep knowledge of data and storage optimization.
Overall, the DoKC findings suggest that Kubernetes has become a foundational platform for data workloads. The focus has shifted from whether databases can run on Kubernetes to how organizations can operate them efficiently, securely, and at scale in AI-driven and distributed environments.
In short, the question is no longer “Should databases run on Kubernetes?” but “Which databases, and where does Kubernetes make the most sense?”
The number of technology professionals surveyed for the Data on Kubernetes report.
The percentage of survey respondents from the technology/software industry.
The percentage of survey respondents from the financial services industry.
The percentage of survey respondents from the healthcare industry.
The percentage of respondents from companies with $250M-$500M revenue.
The percentage of respondents from companies with $1B-$10B revenue.
The percentage of respondents with the role of Developer or Software Engineer.
The percentage of respondents with the role of Manager.
The percentage of respondents with the role of DevOps.
The percentage of respondents with the role of Architect.
The percentage of respondents who are key decision-makers.
The percentage of respondents who directly influence decisions.
The percentage of respondents running databases on Kubernetes, making it the top workload for the fourth consecutive year.
The percentage of respondents running AI/ML workloads on Kubernetes, reflecting rapid growth in adoption.
The percentage of respondents who view vector databases as critical infrastructure for AI workloads.
The percentage of organizations running 50% or more of their data on Kubernetes workloads in production.
The percentage of organizations attributing 11% or more of their revenue to running data on Kubernetes.
The percentage of respondents citing cost optimization as the number one priority for 2025.
The percentage of respondents who view edge computing as essential for their future data strategy.
The percentage of respondents who say real-time data processing is critical for their AI strategy.
The percentage of respondents citing performance optimization as a top operational challenge.
The percentage of respondents citing security and compliance as a top operational challenge.
The percentage of respondents citing talent and skills gaps as a top operational challenge.
The Data on Kubernetes Community (DoKC) was established to support practitioners and share techniques for running data workloads on Kubernetes (source: Data on Kubernetes Community).
Around 70% of surveyed organizations reported running stateful workloads on Kubernetes, signaling an early shift beyond stateless applications (source: Data on Kubernetes survey 2021).
Databases became one of the most common production workloads on Kubernetes, reflecting increased confidence in StatefulSets and persistent storage (source: Data on Kubernetes survey 2022).
Organizations increasingly ran analytics and streaming or messaging systems on Kubernetes, indicating broader adoption of data-intensive workloads (source: Data on Kubernetes survey 2024).
The DoKC 2025 survey shows that running data on Kubernetes is now standard practice, with organizations prioritizing cost optimization, performance tuning, and AI/ML workload support (source: Data on Kubernetes survey 2025).
AI/ML workloads surged in adoption, and a majority of respondents identified vector databases as critical infrastructure for AI applications on Kubernetes (source: Data on Kubernetes survey 2025).
Subscribe to our weekly newsletter Kaptain to receive similar updates for free!
Join other developers and claim your FAUN.dev() account now!
FAUN.dev() is a developer-first platform built with a simple goal: help engineers stay sharp without wasting their time.

FAUN.dev()
@kaptain